var/home/core/zuul-output/0000755000175000017500000000000015157267143014540 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015157273433015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000155717215157273255020304 0ustar corecorevikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gf e?KٸI_翪|mvſFެxۻf+ovpZj>?xI[mEy},fۮWe~7Nû/wb~1;ZxsY~ݳ( 2[$7۫j{Zw鶾z?&~|XLXlN_/:oXx$%X"LADA@@tkޕf{5Wbx=@^J})K3x~JkwI|YowS˷jכ̶]/8 N Rm(of`\r\L>{Jm 0{vRFEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHMZ8HGU=y&|'oZƧe7ΣԟRxxXԨkJ[8 ";ЗH F=y܇sθm@%*'9qvD]9X&;cɻs0I٘]_fy tt('/V/TB/ap+V9g%$P[4D2L'1bЛ]\s΍ic-ܕ4+ޥ^.w[A9/vb֜}>| TXNrdTs>RDPhإek-*듌D[5l2_nH[׫yTNʹ<ws~^B.Ǔg'AS'E`hmsJU # DuT%ZPt_WďPv`9 C|mRj)CMitmu׀s9r;:J_T{*T\hVQxi0LZD T{ /WHc&)_`i=į`PÝr JovJw`纪}PSSii4wT (Dnm_`c46A>hPr0ιӦ q:Np8>R'8::8g'h"M{qd 㦿GGk\(Rh07uB^WrN_Ŏ6W>Bߔ)bQ) <4G0 C.iTEZ{(¥:-³xlՐ0A_Fݗw)(c>bugbǎ\J;tf*H7(?PЃkLM)}?=XkLd. yK>"dgӦ{ qke5@eTR BgT9(TڢKBEV*DDQ$3gFfThmIjh}iL;R:7A}Ss8ҧ ΁eor(Ё^g׬JyU{v3Fxlţ@U5$&~ay\CJ68?%tS KK3,87'T`ɻaNhIcn#T[2XDRcm0TJ#r)٧4!)'qϷכrTMiHe1[7c(+!C[KԹҤ 0q;;xG'ʐƭ5J; 6M^ CL3EQXy0Hy[``Xm635o,j&X}6$=}0vJ{*.Jw *nacԇ&~hb[nӉ>'݌6od NN&DǭZrb5Iffe6Rh&C4F;D3T\[ bk5̕@UFB1/ z/}KXg%q3Ifq CXReQP2$TbgK ء#AZ9 K>UHkZ;oﴍ8MEDa3[p1>m`XYB[9% E*:`cBCIqC(1&b f]fNhdQvݸCVA/e.# Okx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'}׃nxuoỴRZ&Yzbm ]) %1(Y^9{q"4e?x+ [Vz;E|d1&ږ/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*t{zƝ$;ȾjHim @tBODɆj>0st\t@HTu( v e`H*1aK`3CmF1K>*Mk{_'֜dN${OT-n,'}6ȴ .#Sqη9]5zoX#ZVOy4%-Lq6dACYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tuw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!fޙ-did˥]5]5᪩QJlyIPEQZȰ<' I,($F{ձ7*Oy 6EK( EF #31J8mN .TTF9㕴/5~RxCe,&v3,JE- ZF5%Da,Gܠ*qI@qlG6s푻jÝ$ >8ȕ$eZ1j[h0SH,qf<"${/ksBK}xnwDb%M6:K<~̓9*u᛹Q{FЖt~6S#G1(zr6<ߜ!?U\(0EmG4 4c~J~]ps/9܎ms4gZY-07`-Id,9õ԰t+-b[uemNi_󈛥^g+!SKq<>78NBx;c4<ニ)H .Pd^cR^p_G+E--ۥ_F]a|v@|3p%kzh|k*BBRib\J3Yn|뇱[FfP%M:<`pz?]6laz5`ZQs{>3ư_o%oU׆]YLz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd쿽d@eU XyX2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^oHmmU̸nG pHKZ{{Qo}i¿Xc\]e1e,5`te.5Hhao<[50wMUF􀍠PV?Yg"ź)\3mf|ܔMUiU|Ym! #'ukMmQ9Blm]TO1ba.XW x6ܠ9[v35H;-]Um4mMrW-k#~fؤϋu_j*^Wj^qM `-Pk.@%=X#|ۡb1lKcj$׋bKv[~"N jS4HOkeF3LPyi︅iWk! cAnxu6<7cp?WN $?X3l(?  'Z! ,Z.maO_Bk/m~ޖ(<qRfR"Au\PmLZ"twpuJ` mvf+T!6Ѓjw1ncuwo':o gSPC=]U҅yY9 &K<-na'Xk,P4+`Þ/lX/bjFO.= w ?>ȑ3n߿z,t s5Z/ Clo-` z?a~b mzkC zFȏ>1k*Dls6vP9hS  ehC.3 @6ijvUuBY hBnb[ Fr#D7ćlA!:X lYE>#0JvʈɌ|\u,'Y˲.,;oOwoj-25Hݻ7 li0bSlbw=IsxhRbd+I]Y]JP}@.供SЃ??w w@KvKts[TSa /ZaDžPAEư07>~w3n:U/.P珀Yaٳ5Ʈ]խ4 ~fh.8C>n@T%4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F on~$dƹɥO"dޢt|BpYqc@P`ڄj҆anCѢMU sf`Yɇك]@Rɯ?ٽf? ntպ$ˣ>TDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d cRKw*#zJ9tT :<XK*ɤwoJarExfKB4t@y[6OO6qDfEz]1,ʹB֒H ֱw;SpM8hGG&ƫEJި_1N`Ac2 GP)"nD&D #-aGoz%<ѡh (jF9L`fMN]eʮ"3_q7:.rRGT;}:֪a$)gPSj0j3hLư/7:D-F۶c}87uixoxG+5EekV{:_d* |a%ĉUHSR0=>u)oQCC;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKdĹgJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?G(" 뭗R==9!nKErHc1FYbQ F;v?ob-ڈFalG*rEX}HAP'Hҷ$qM9(AHx!AF 26qxCdP!NZgҽ9l*(H Žڒ;̼|%D Ɖ`Pj . ֈ,ixp`ttOKBDޙ''aLA2s0(G2E<I:xsB.ȼ*d42I:<ŋu#~us{dW<2~sQ37.&lOľu74c?MՏړ@ -N*CB=i3,qjGkUտu6k Cb8hs&sM@-=X(i7=@He%ISd$&iA|i MiʏݸT{r[j顒x.Ƞ"m@Hy_I )j|s#RLyL B EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)@kF;'ta%*xU㔸,A|@WJfVP6`ڼ3qY.[U BTR0u$$hG$0NpF]\ݗe$?# #:001w<{{B\rhGg JGIެE.:zYrY{*2lVǻXEB6;5NE#eb3aīNLd&@yz\?))H;h\ߍ5S&(w9Z,K44|<#EkqTkOtW]﮶f=.*LD6%#-tңx%>MZ'0-bB$ !)6@I<#`L8턻r\Kuz*]}%b<$$^LJ<\HGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dXV0e[g#B4x╙✑3'-i{SEȢbK6}{Ⱥi!ma0o xI0&" 9cT)0ߢ5ڦ==!LgdJΆmΉO]T"DĊKٙ@qP,i Nl:6'5R.j,&tK*iOFsk6[E__0pw=͠qj@o5iX0v\fk= ;H J/,t%Rwó^;n1z"8 P޿[V!ye]VZRԾ|“qNpѓVZD2"VN-m2do9 'H*IM}J ZaG%qn*WE^k1v3ڣjm7>ƽl' ,Τ9)%@ wl42iG.y3bBA{pR A ?IEY ?|-nz#}~f ‰dŷ=ɀ,m7VyIwGHέ 2tޞߛM{FL\#a s.3\}*=#uL#]  GE|FKi3&,ۓxmF͉lG$mN$!;ߑl5O$}D~5| 01 S?tq6cl]M[I5'ոfiҞ:Z YՑ"jyKWk^dd@U_a4/vvV qHMI{+']1m]<$*YP7g# s!8!ߐ>'4k7/KwΦθW'?~>x0_>9Hhs%y{#iUI[Gzďx7OnuKRv'm;/~n-KI`5-'YݦD-!+Y򼤙&m^YAKC˴vҢ]+X`iDf?U7_nMBLϸY&0Ro6Qžl+nݷ" 㬙g|ӱFB@qNx^eCSW3\ZSA !c/!b"'9k I S2=bgj쯏W?=`}H0--VV#YmKW^[?R$+ +cU )?wW@!j-gw2ŝl1!iaI%~`{Tռl>~,?5D K\gd(ZH8@x~5w.4\h(`dc)}1Kqi4~'p!;_V>&M!s}FDͳ֧0O*Vr/tdQu!4YhdqT nXeb|Ivż7>! &ĊL:}3*8&6f5 %>~R݄}WgѨ@OĹCtWai4AY!XH _pw騋[b[%/d>. !Df~;)(Oy )r#.<]]i-*ػ-f24qlT1  jL>1qY|\䛧\|r>Ch}Ϊ=jnk?p ^C8"M#Eޑ-5@f,|Ά(Շ*(XCK*"pXR[كrq IH!6=Ocnи%G"|ڔ^kПy׏<:n:!d#[7>^.hd/}ӾP'k2MؤYy/{!ca /^wT j˚ب|MLE7Ee/I lu//j8MoGqdDt^_Y\-8!ד|$@D.ݮl`p48io^.š{_f>O)J=iwwӑ؇n-i3,1׿5'odۆ3(h>1UW蚍R$W>sngir^$W v:?_ͬ5kݰw[!$s׭dֲcUh=Ɩ9b&2} -/f;M.~dhÓ5¨LIa6PnzɗBQiG'CXt!*<0U-(qc;}*CiKe@p&Em&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S [;Zcq6xMY 6seAU9c>Xf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O wmLeL6q28Fdvj5 )YY$Itt'4q(@pɖ9Td զheqJ/4?T5(߄*?T]xD ʀwdYq%7wrl!CUJ&!\,`R8XŨy4eUq{\~)ϣB6IPB3j2H`xTɄt U=EySF?TUif7W"6-PQA6 bya40O]Ă E`w'Yd2xAmǮn21O0Z/H7sI H"C_I ^Pt.Lcf{혿Dh&EYU,PބX oWYu(sB?ir x/>`޾qx?ҩLڮMH~]yT x/=U-Y~V@;Rd2A":(Kxߤ/Չs,Q=q"*<QG=2"<%,T"$V ^z1rq?6Rbch;x̻Re3cr]G1[ZDx#Z&e̲X=bs FO#A|4Y˄+(a$<qւ̪H[zR<oam~Kld+֌7N!cP9(2x-WjT𒧠X娚M#%EG\֨*h+a~GS)G.ד;梚Լ&Sg$U.2WT'(/G)!vFt2o)|pEaڻb<D]b ws}uΛ?']\0}q.eTu)yzgP`g B `8?/D7%wP|R[q@݁@AU gm4v\%8@dw}/+! EhkW/] fw=yy kl Gavχ?eiR 2<9z=6@\^W嗸Q9Oߣ8~itM@@$.hu٨2`U=@xn"RsՍ e +k剂p\w/m.6Qi1}9i|OXTz tԚ#k}6]*WOxy׶㤔m*ΓG#A_(IY<˾{ {G02݂LO2tq]YVyS p.ݫOˀNf4yQ8cn?T',PZb\H`\sdžSvg7?;p8KLo4eڑћO0 ׾ X9-(cV6/L7Ɖ ,YPWRǑMV4]ሯz/?%ъFrwGdw0,yN^IQbVar6u' aNԷ.JUï1o$YMph\ V0.V1Yba[9>^FHM𭭈PlnC2_jw`I Pi:^ćAte,UIۆVԘnYtfv-jv5W=eq`ُ8 (ReNL .*/onh?nO9- 5E.:`6.K﫢3w-'R{oOD4{pض:ꍵVB#㩬 .\zm)":g!Bu9ߢ1DT_UtqDKP98ۜ(Ņ6i\쮹9>ɮo=*ugPn kklj%,3Վ[ju ɆJRPl ʘsuOs9D2qu" ]܊lwt>ue$ id[U55P_&\::v*:6Ծ*:@r"8 __= eě>FEW"E< gnKLb߿8V=lKu5[JpSJ.yV{=ʔǬ-a{;<[5315,2mY8l~an]T):ˮյ'Fmd;=K]g ~"CvbE1. RՕȓ|v32GHs $_\m9Z%Q=Sg(Jk.Ui g=poXg͍lG&j-TD 3 >ݞP|w'3zfdOf1b=`rS={b<(an*tF"grd$#ww2Q9\,F㹾ŕ u r" [Fc)ŵጚ\atft1Jɞq`{~U:֞ b[~mF=OdޏQQkOf1-!f 7Oc)#݋֌!nj*0c;Ŀ7ۏQ81V3هU,`a11 1jl22! vl\x]2t 9YYM`b(Ұ󜙅V.;m Bx2o)|p`!&Z|ybB/K@OokͿ!6!ɴ{W@< p,c2.q{S-X6;53rA)$ &@&A j xq4a(Wm1m`>N2lY) V,6"RVyD? W;bw<P4/Z {:kP=F,pkL$ Ҫ]=|#>j[> 8pX]9<ŕt`׈}h/0B,Da `kv`Sxc8PЭuPtܭpX46-3Y|{Akvt6n' 1x*s7xLjQkzk\`dÇ "rr}[] ^$<@w$Z /|6M#8}Ja{4i/dZW}`C}oGXr" @. OךW y :ǭ Oi'gpF\5f]v΅D-vU߾ia鰸[4ИڌAJQ`9] V3F(x Mݖ{&m cXܥ \0Jpӷ28,[bT#ZxedTpx (^3BJC`4JIt^61 2}mzᵶD>dLB6Ccncv8 )1wm~*p:L`LC.?g_TwHw7x鵭04+[Zfm -YFސ' A: Y}4е|%Դ㚏ĭ-L|#":C!OIwtْiߡbv\ Ug`5 CrH)\Fڬ vzy iZٖgXsӡif .Y |HwVDNJ̟"ˋgO`w:DQ/6ݨgZOzpD\X=sӃyzfIځ7dȆP|&b`{In}=PCFK)@~5 'k)~/ʐIçiB|0LG"B4,k. fUA.v,л O u냬V2!\XOeP@%꧹zIQ{s8⽣a>GK]X^K|Oo|)/(Zcy:63}צ͍:Owq|!>"é<7v(^<.n\JZez.{Mè; w[M}Ç,uAĶ=bgVCr#/NcrpDs`~}q"4 ižvϲٳ! Q~:T Js@ >HLM@H Ǘ𚾆z#> MsT ŇJ iTZ* ,(13c,D/x4)QѳZN$P֢0u}7p-+%fOPIKW$-qy?+У M! 'l8eVWK1$O- p}ZUgWQW?E9G$mPbhhB(NbVt84GDmGA=wT=2n}[-~o {W×F!HR &ocʫ@mi@LR'ۋa1!| g<d,w+>H&e:SQgI>dlhrSaWjL, XsrLL2\qH< Q&E-S=)d@ua/MƒC%hv)PʆGVK9sAGjL/3dxEv] Y "Q_YW3KBa i-VNVUG9KNR7Q`0*JzXPDS-H&mScD 56fSy~i ڢ|ZGF}sX:!s-5S.R9]P 02kP.ao|3ַ$C 2\^Ui(JoգmLnTq*"5yv>4 Iy-ǨE/D+M :3ŦHn?=|M[lwt̕c _|G+}ѿ1zD b"vwLmwF@m8RBW=ꬺ+Ox{Jld"ͩP #%^pQpU0dK xؽ0 'q-U`HlpVh3Y21,Y?}3@Lnbtۜqz`:kVl*-,Lq&{<9fho*2WxPAiZ!̖yա|틣KJrC\A$;wx7H($U[s$ΟB-\ypC\h(U#Ry1#iyCMP5ߔj1[Ϙ&EmBӒ8NUFE"7^UxH5ATxȷatLZ/f[; 9/Fjx.3#Nefp2[,4U/cw?5=a~BQ M ƺRz/HP_y6o8@[s_(6'(URI+ؽ  ydy,;o~ͮop7b9 wNy.!~_][or+UQ/A8 WⒻ]T^8eOz4U]j}Y?z''_1-V>\y.^KXƾ[kmM :8M$o"){/9i6<}BK9qY2=mfVQ~W@i;_HP{ӿZ7{㜞~Y[fj2f_:sffߏ8&ouJ4úq sUgdž:6Eؼ{ Nqej~)O?^xT@crVmmvVj[l}1"<8{P~>o߇]#iHrM G6K}|s4E[-UCR$4sTL .TS2cpK)bQCJbpJp@{o&iW4?_Ul(c] t,kAY( m {$(6Dg 24)IF|c-mIHJ#sCE;h5xx6 {ϕI[* v-@PQi|("U]^&De~I.K;3}TV5Q$vAPEPIͼؤ*Q74M*reAhlv(}WCYi-(u $YZ)`:VaPExQ۬\X)E'<_~ziJ_P"z!fRx*z} *36)S2Jw}rPsNd4HttܵI8~gG-'1fxZ8E`Cro:4HtRHɚfGrdsdưtuh^Xu>#7킣㋪J┚ک0lA0-Oo'Q% /p)bzY;JȲrtxd$OZ5G&XCgLo$.ڀQ`hMAG BX{AI01wt֍uhdH3,D=D4[YfF)b`vbJ<6wnpT2.Ms` uڱަ]|uEIE}/D"^H͹I=Lַ}3}I[ 9FTB=lΆ\<8 nl&VF D?W1`އPI8ϫն~pEˡpJ,18s'<+u(%M-$K n6N$fG5v9ۢ]^!{砸|!B̊D 8kUIc߭ӝ7(Hv'2{}vE&F#Sᤗ;c28ys1 ?0A%&).㧏NtQM)[ >F#ت.3"IXN2cJZ(lBc(8 5xVe"$8د1drES?6Q(W-2 _K\-^^s vlDPaxX&$ L4=4.neJwM \5(tO4FL! GHɅAĢXf4 >l>qtRHm=Xך_mM{V!u0"[bŝ"ǽv1v"vfb#el57.8ek/o|\; z"u v-sn6XɓR%OW4\-I}n0gɵ2 .0kL]7Z Mz|_5<}쁣bQa.|XW~Q}33nJg4Tk/ zLEzg(a/o6.8{o`l0Z*]5x,P ArS2&Oď"/kbILFbPT0!-Q`219Htqlgnb9NSְtNP{>8A=xPvD7]yz`yHoF3L>O 3G8I/o3 vmCRV+,G+S110B*ݽ*-)*\k.f!JN!֦}k)FXwfڂ \AVߗ*)e)f<c} WVvkC/%ŗ +4͗|Y[S@(WnY0kwzRDl~;?8J~Gz&qO! iiCSq𜷍M9H֫i݁V F-+D" m "~hOݾ IY\@!qO{҆Px17*qe*Q2ʳ&w vqpw-+ЈL^/=)GJU.8JP&8xu3)7:?.8e_[87*}kوV뛯$L]ly{vlVb(Dfa\*DFTr!Tn'.F]ʓ=ZjN!ď] <U>zhSE7..`3Iy8{FqnRoW/K V1oghPUܩ\A֬U xk iWֲv@1)\h\J@B43آN |iZ%GצyP˳9]hz]k]C ])n֭֠PlJx Z˶SHrjES+m8e@ R߇0v7u7?DӇ40Cj-o ) ^r]tSGWw!+nTuELd h)nY.q{Ҫ&U"uQF&H4]Fj ΈU(k _>d}يMvwxqu>ٚ9!^#l0ٵ;Io<ɻowϥ땉1z5B`eN3Ŭ݊BыNqmOUiz7>muDj gۼKd*$,SKInޢ]%ֳ֩3$L?!UxL6T3e 4صY?O{g[x퀷coǷͲ/#vm+(8ش=qr]نَb"n]ـ(LuS F`14d4q ]u]vat+8Rڪpnv' Ar]l8HB6+u|gzD@?}tb4N`pہ&BT.?`ElzG g?Ie0# ą0'MAČ^;j""_%X\߀L0̥OpT)'hbi|/ m34|忝o&ni}|@ZJ%;}zr70Ge$ dm^foO;7Rhw:}GY$PF݃x}}ߏA6L/h||6R5ŏ G]"+GY8Z%Gȓ< C;v>ݰq\=`8{icIR(G,m Lʁ]PJ,bf=ONdy[(@'I0@.q|Ss5_'ї$=W6*ԉ`/h}(؎Ͽel| p2uN%c{PuGZ #_ܻ֓*. +ƟW&h4Mմ۫] m%0UV+]%S=mv2Eph/+>>^y`>h$ []JYvXs&3moqfz#o\O.(z%iq]T,0cةY .\8 \@TEr3-1kLs|cFǽ,=*qF~1WpU*ߧUA1^y6pp8)ד^Y#zvWF NU&%v`*Zx"z DXi9O[ x><*Rx&b-/#xKCWmKk_wx[_> v]A) T[j:_v5[ƽڜq*az6Hp^7̊EXy*W&s`Ȑf$zW\L"\ݏ/AEeACght$n:?|s25$ӔF\d:}Ntkw3sYd ߼7c1apBseTX3b @.H05>P0O&[_ |R߿-,B*TM31e7χ*"4#^%R) 5 yLײVLlvRx"kmwb5!,ݟePf^*.d>R#ۛ_ }i8_D^`t~q8@c[Se"Z¹&B~{h7' JD|Y f_7>y 0?.ɰbOgSQJQO`zjUZBUɈaM}٫DzW`_TV)Y ) ͓䧂jcޑ?"%Sc ;)/1I]ƕI w+u ۂo}|HlnIRCs(U+O ,+Pok5 S{e0.E)"tAx:e k(j2FK1+ (;g#-/$]͈aL vC81&*£،rP~pg9%ߟ^8>Nև`!1+R/")60DcTJ Li'wѐ&ﻎ>NcÙnC'P g%Q+&6w+¨Dkii!E3Ű+ \i,  B9!]\ƈ56T-V@L)1kK)D(XJ%CviRMڠm6hF E,cRScI!Z))^*Xⅵ l kX+lmvcPl]1`&m&^Y)l j-W M 1MamPL$kxr3 TKju16A+]WUQ{L&kj6 b-$.'Q%I>,AgZCH^C~]SaDylmoB>B<*;LnKFpESr 1=@3D&,!pCТMjr1b6I}uT]'N|]%@-FH! rio\ {8ǩa&ox+a| h*yV E/eGqT\F 3 i[*qB~|2n-_.([n3WCc}ק=Swwo{4V'3>ad'L@L΄NTS42q6Dk-{+" ŸT,#~ ~*0]Yip,lW.,]ZYfZYnV6EZ&i>Er aul*㺢y_mu2g8n/_ ggSsOs=P%f z.l˥QyPZ-A%q(C)_RKzAJ+ROcvK.B܋3"t1:08ޙEp9WiD+Ơ]qzG7{= B?|dAuN/~P0'\fGo*A{|p1GhG3[g|ign/>JuzĆ[ (9gܖt+47bZӵq4KTJD3pO$w$¥BIp8@go@N \ubGXjYwPHuyq3#6:d4I|8d]$nt?aUfq۬`#=g9"*1dozWɲ"q@aMJ< czt90&g O5<=ʭj8D>G5rfxyk%c5-|MQ >uScv:.דI!Fr?dUH_p) &G I'y{;ԋVJd|Z0U>I=FQ~b_&عY4T .EG|r  >@ %t_޼e2j~٧d|a|^b4FM+_f~: I>.K>q>N?M(;8tvq3ބC p|i?< _ t:rNr]F΂x䖮מ#Qw+gYWX@C3gtTW~lұD͕V}glWe`ۺ|':Ps[5Kj=6Yoԯ*Q]׷v. O8Yg٧QKwEn9J_e:JO''4a`wR\Uـ1/73y9?._o}9G_ξv4WTuq[r>GW~ɮޏ `TQw?/,.;_OB%WFf2 ]n1Onhu Ύ̆hslyyq0,&ga=Zj#iA8WűB~CbS.{[ʠ-~{}亜heΉyAkRP;!tIGC@(hLA ^9+#ye"ålX1M@3}q:4Ro}w(@H:wt%h7 ~lus9|eAQG.xSu4q+flNVf896[AQ͑T# (dڽ#k.H0uHM}8FkKLKBjDž:'+%X⧢ sOuAМ2rSpKjvvfGk,GLzlaed// S[ͺ}:&.*~jdEKo۪X;o r zJe==t(!u9pS|o!G _$vNje'jl}};Uz_cBpON GӑѮ'k#1MVTo$#  &0hUhpNBA+ZήG{^`<֟9jKoF7dpFÙ2g\{S:.+R˥^ڊ S*έι(2se'~w>[fÎ}M.{^ؒ_n:^-q$ ۹ uM6mfPs68ǥ z/ zǂQE$O!\7tUֈ&?փ$@[6Oz q[[nI}lo~-pPztƘF4P5}E3~8Ú|OhAHKGJ3I:NJN3+:+:#iYQ/fH@ՆP WZ݀ƇRRhYVFIYY#TB1PJVUG3JR$PQ` qkH]*F]Zw:L~CQ4zZG; L\rk`Tb#Bk&  2ވ3#rZ6GRUQxx$ bi+V崡˒` /ˉMD/ hϤCu=qRz P}:+U4jY E͎bp>ľa喷>ǎ6{Z=Q@)1NK~GR $Ҟoj}~m摵|V{J]GXg+ﳓgd'֍3o껧 ɣ7`~WlmvZDsVYk#+y])-Jinh<؅Ji-%>a}WJt$}t*ydz4kaMjaJ(-b(0 bD2qRiw1$nECTGˍ%I˘M]Jr'{Nqolhy\C>0Od' RyoHpgSІ+--D<Ԫq~rk'G(#oohfE@FfJDd`C`L[A!B]JV_'" 20°idEC./}ۂv(7YABLSE٭TL"pxQ4!#$P*cwh> !(dG_3 O*M7{?\RLK'wd'?ioZ>^߶UgWVF@Oԣ'x:RY-d mץFV`P)U`jR{ 69IaAOLmH~RGRԑ:0iO(h'u:28-k'u=uFvz_ $)8#۸A)8ZDNXVӀR!,n? 9tWw즗5/pF|wnm*ߥ W:g+u޶*u2YUlKd'N*=e$geeزjK6g8Mи~|ƌcc.EWïzs{ V`Nu{,/Ž:չq釅/ɋ1s Xm.CޅofNqTͥz]=}z ﲶ¼NہJ s`<+qyo@hi~;۟h\3QA:vzvE^~6K\Oɬ{u]EWz@fҏߥu_sw"o4QwD7[o2)}(.x6.N+B%ߣwwߢ}O7_OMG"7BQcu6|}E<̚|H4?|bK={80O,s1Q̶l+Le;~x?UH}cޖNXPz#0 >9|xؓ9|xtO ϩ J^J267&6|[ʞ4D~Hzys[@ QLj؍$i&CFʑ*I 5T$c\Aj!9*8c38c9xHF'WZ㌑b|fj`]XǛOSW'=O7ZEӭn 5P7I]!sh]|e1H9jtP5k\Pû}e^CQT(UD֋5MZ/wKYT!IHl&nJ^]9*%U"PiAO&MqJu%iTT:''T P<7W:F8S2EȔ ^ozOH"(^TVT^e+*Y$<6KUpsV^ef|JW2/w6=3jG  Zm ؋G,D7Q V{ÝѦwXq+m!;gW\!&D*H)܄-h!eJ &ojPNn&zq]TwnzFslJU\[(t.*mXp" V}Gm.pyn#{[JSu@5gPp/)B;(JJGǃ^#,wCpc[*'i8&Zր|&ykuRVRSEo#g1*F`AOTq}Anua9=O8=OdzmҚC jqAaoxUtx8 wɪa38ҴN4CFNCq6vy͓[-{.?:op/fyxF,6>}gKKvBc 1y x"E^v*^ g)qBEKs0uAmEؒF`OF.;s^8j!\5̗@mO#?rM.?O˗cP4ָ,P)j *l?R5۵wkW,#}@?V-Zr)KFʑ~Mmo]AkA =#X[3HcVT3o3**yopjmյyHīhqqYY5R(Ca  5Az<.˫ƅO$fRɪ)lVOuCk䤏Dspj:͓rU|k#kmWrP>F# RZi! #9Ԝ9MJ' kNsٜ|MY+Kxһ•c@lK2 XY2KR_Y,K}!JȥHdH9,!F] 6T.;|eVwir6q/(.VFʁZs?ԭ}ђ^G@]t$l50X0>a 4`Q8k0MKԬb0Lg^(HPjqfLss p$=R=!%Q 9<:zw[L]S;vW[ őQBx)7>3Y4ڸ~jϾM.+kqQmppW&6(!mQs0nq<(hemM0Za@X۞)@m-W>žfC#ދe(ٜNrVh< m=Uh Rxzr ΏHw16 ZF8`SC-T λPо=*K{5dl-cDE8˫0R r&ZR @?T}]|YA ,dg9[FV5Hf-]:RmAE3R5 9ey lp?8_]E\Ս@b Ktx!FA25뜗! uc"b}g8k⼍rUl$ׄ~ ןBS3.Aoub,R)UL\͸Lg\:6je3ʸPV` 2kQ:Nˣf1()VFIYq^s2+FDǪb}|bd1@AC=żb@M b+(L ۛxje2.@ǔaT3:H9Jg; қfйς|wұ8&Zր|4!i԰5GgL<:.)SfSNQ auoy[9Cwl:<&G1r*cxPA ڸr0P߫rP#GG(\bH1 ~M2 Cj̒dշED(uHIe|H9"ي1bcU"ܣ"iS)Fي }kܹxwfG8hю21OgMN!H9Lgt[y]L`J^4_sx!Ƿr*nk0ݱ1umMb& NC_ B]E$8j ^B#xD1erxUG}*߇S]u'U= d-`Vf1֟o=phz۶l*XKf;.!iVl9bJ9Ig6,)i`TG%5&p9rl#hE1rƂ[ן^;m?߿1C죏؆` " w N!2M?{xpc$YX5R W|Nw\as5TG9dw/3%4bםud)GZp:`*gݵ{r%tw1ltR &u䝱P Q_?$n(9YFʁgg*䆚NJu ) c"#R0R-ų=_)QC򔔳e@d7xQ9R7R2̏ύ"NOA3> ?g0'ŭ7G l:DzʲP[2_zv|&ESpeG%d'V*dIaC`TZu E axuXA[=)}Xpߗ(43@nJ$z}T=qM$[xI Fn ybvN㾯:pgzvo&m{ ~{ ~{ ~qюȢ 5Jì <9Rɱv6; )=ֺ=z@( QtP~us"ߐPC__< -t} . ZyF%#oE4ibZn-Gd1];>2ߗ6$ !V)E3}*~@KRQ\"^_ Q Jz|RP9Vh+NTujo[aSrA,Ly:1DZ0ſՂO?b{_õt ?|V' ޻0q 7gEy9A_!,e\o-GT!*vR^Xp UeS jl}"RT`¶CjR).=Zf '°[\;t>) Yw.K 4SD028RXjH&]|cr9.V,{"#G F~;Ky>:SrXxRQ5ե.mu9R\P3rֆHf7))ʴ5 CM$$6V 8e9MS AqoNdauE MIҿRzؿTI"+*!YoVNz>X)GJN>\`wܿA*)Z߷OW9gHqȇi! uF}S֭%Ӓ-ߖ|(nCZ*õRR.=#僉Ԥ%mgIYvIϒP UD3 H8hIR5²$m'IIvt &RQ|jdyemel+cd 3B)#9?=ofťTJ-X<7S9\=XK0L$0eӭx_߼YOA~ gy'8*GQ r쥟w<+xđ./u!/\΁Fc=3]{#t+v.f~2,맞F ;p5ݛC+w{;n n:vqFw?E{I"Fa^rR ×-b mIZG*lSZuxA¡tTmHjD^=+o>"Q݇W(ΓKwg ­XBiV"Qv5!mDقs~4pWKžծvwakPĮ]]xk#X6@#?V=|3EΗa>.@Wq>kl(m{H ԸS HysQpcCߔAQP+%ẋ]YAD5$Q|CzRa10*$g,SsPĖ,zRnY[j:֦#JlKc5|TN_ppȂGQwW5R1%jE"C !hep{3}(׈YWqw.--o1.?C,gmt;p~&?D߽?[ Ĭ4Po?< 4Ѕ>BmUtEWHz>{$;x1K;sBpN?=B:Ědiy׆Tewnz!NM'&,s.(LPʱF\%BSJv٤:Z3r:4 ,$2ZaN*'4i*xq05NInePk5Y٠:75PK{(m#Z.gvDk1m3Gu)NO>甜S{M{KZM~ 4$f?ά 3 ȈT&&rN$kp.E(C5P^G-0ZٛO74XX Vʹ(пʩl`tļqhͷKn 6&O&,7.GYgK -y#㤢HX'!NvN;q4".A\5߬YcDCiLm<%i8ԁKͦZh1[^P?%5ZOA6հ'.^@O>((gV0#lҊΦZh(a8[y d:XXCAX&C5kpւQڮ:Z?D_z].mx ?SX5ܯ>+.z~mj-Vq0x,~#84B)NshW>w sߵStfwu!$em -FI6"ZbaZqZ @nl(ԁɩƍZYKֆM[RI-V5Y аY*>y_f]o꒍M5+-G-Uز*Q+dX&(B&I-'ZlFmb1b,-T7d.ke$S!zSq<=H\suoV*+e(N(;%KWC޳sہyCSDpʗƠixy r?xphzozߑ0G2a_MP Py4g ^rӂ0*Y}ua4耡w'}3ؘhstSixFEKvUi}Nz𼹺L\=]zҤD=Z" :09cͫzlX%ׂQ*Z!8/f_o^^/Oa/ěZHîOE/a?.zy%HR{{.h;K#ʲTPbaX.;x4+q-w\u:DϓԬ{>3ރ>ʄb:Zd " )PJS'D`D)\w +lZ ( Saap퓁7vi>2G{Gq#RËqi -M+00 ,dBB^,J&gˊS?\I\uAtrΊxC< !@{䗰bi(@) %.5W`vKp ۾W_I1R`j?AO|8O /m[c #roo^/Eẓ{wadYCͮFŜgx^<E޾痢~>&iw!^߅:ۉ/xO'(&D{c\ 8=.:Żˋi^__N녿|݁SmlfU.YOOؼkt"|-/f0AkE% +o~w e9]5Q l~]JN|~³̽I/!Y /+(7j,t mOæyrGc΀ш]Ȱ|cTA /Y]7E/2J WBQk8$"~9qؿȿD(E epI)\r2)!(/o m@L'Xca1F 118ʌrFqH $}&"h,E!Z21"jR޵6rcٿ"amlvۘE&@/ >c!dXr }-ɲQԮ c vK%)xH HԂe,B"(.&%X "!#2uH*nF)K<;Э1  : 1LLAd $UMSB`iT66@% q_e `I p2$VJ!(*挚10J8 tseԸQP5+b#JQ @+drABAjԃ2ZFSQ랔u$$E}^R6Ԉ &bq3:EB}UTXQx %Ŕ`h#"HS-C^B$!Z댉0Nd3?/t[VK<#ŲD݉s!u@} ]c?uwJz=zղAw!T8MuY = 7 ^x1Hu`KyY=X iT28Ut)\]Hh)bxLqΠ&qB|IBBGą1 @P$bDM W$2+BJ*{4m2A)حL( 2pG[S`Lt Vϓ#V4+JXm))F aH+L7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tΫ5HkɨcGvWoJ)QhQGuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQFgC<;݊N[LPa{y<};7fTJ^mb"y'?0O}Yf _7zv51-24\-$pL/KM]<9LIwD c~_?t[7v~/g_ކ.7OtYL1WtyrOOS7cYOϥ?r2|W|M?"-9j`i $޿9˸eϝ5C2tTf;;luw}n W./f_e5qXΘFr+gB4VZX+`URFj5gV{ X=j}%`uL7=rT0o,XZJFJ2ƛF*'L3`)̍5N#SZo% ޶ K~ 4{Jm7yZJ G++ϐr{ZksܹGU(Jho<ƃFZo]3-뼓`ǝ/՞`gO6V0fsףULfZ0đ)`oYu3a@kkXό 'X}Ƙ5J53k]+`sV2gobjXZFk9wN4qY:,#O4{5gKs˹]0kvr?Xɵ>"Fjn{U^7XA#`-ShVQ gvk,j X|>lva91_~_]m-]"~p4Ď> {ApO~9GKj J:v~2eK e"IEӴ7Vۦwɶ%n*nA ϶'?\KçTB=|~/Opߝ oS]69S}[j:v3 ԝN\[)󴘟㯳|~6C4-nwP-'2ahU_N~c9aMXՖ?a)@UFPv QY7y> kQ12bjš>\D}$rwNW-ew Ⱥ7Fi쎾.mw{ |o8W i~7rvǙ^:6_;CKϾwOݘ+>Yw{m%zjgwj Ƿ Uax l~a(_ <|aޯޟ,#uEכlO#7Ϳg|fa_Bi%KAJ]-8}R£4<] $,e"ƶĦXv1ۖ(QOy^M-p+ˢZl4*y`6w9Z/b"Yr4!p-Y Mhm1-]JZMH{ ;>W6״|-%UU`gk,V X6BnhCw+^8exRd掎R29fxT\Cu4C(Z4Kuh ٻ8WzJ,U!E@\0k)1 wkϩY,/$^NЃ왞SzL튥ޙK:Yi1&3.cf݈fhi6'5*KN }N^{{"Z#G]mY;NYW_W6JIk;sw 0+K"LEt(ιKhf>|BȥBcVM{'(\~B>:D!?=.tƢGtm.Si8SѦřJhQ1'v!g55(9̪͝j35Ժg9ՐJJW[I3՝s2J u}6JK1)n ƘtjRhK si 'yf̷)H[eWDiǃS.RHQՒ$DH,8dky3, ^]"FUi'X'=h^,A.>]0dTW)鞺50sS`w"R VIy:ZnB BQw_Kͺ#oG0L#mTȗ*\$0ANJA k\^ TTlPtA[w Z+=xC9v+aD)Qn첫2X ytp %m\cqeQ\Zlw:vk]=^1^-FU-e=ac!sVQmPMAJkGU((JUMA'Sv6qA1c85p\ jX&EfC2!՝l u(\!G!hTPk@oJ'm4dLBX,8IH&;j =-V(!Ȯxd'uu++ȸALAAXw[(0J@T:2];Xi T gf(@Hq;A(6Jm<;"JPAw/uZ~ 8Su2R*.Q>() E\}JuzFy|7 !1)IQ]vJDFHՃ6MVO"b C$+aMh#ǻUzh2&2AyA 5ڛQ̈KUUNBIŘ@Qb󪈴t8AZhGN`߻q &^X7&?zz|϶H3oh"X 3 xs:pnP8B6` J`#eIW=$ B+UX$ETy _6 5dmE@vdm^ ,:P% 9ՠjK0-"V1X52Z!VM;jY!=b8h͛.f _-d pwv}ZKk2} Xc;E% d@e'ps X;q4z&`-S3K.{= ym2Ϋ=7`#[fc(sՕ@.67m\]: >]|V1 ٹ8Wgm޵iOrr@ԣ05N?lHߞC\H6x}lƹろ{'[x&``5 ^n,yA,{54{Es`Ej&`QG˟%?]o%f,r{ޥGh;&"n^.~>jwT{5 yX\# X s q0|X \6. sl Xda?*>Q XSa&`eKRTL2䬞 X ٹ&lb ([fODEYY X8&:LW3 m,`\b Ji5~;.= sQYuVq.`I,+v3\!s ߀eKo$kto>wZWߵ'"m69M{ L k۵p;j׎eeu٦]dOޭ~n~|>bylƓJU*R٘Z\"NNj\(^naWEGz/0[B.W'\wvc[BX(2x?.o./w{yl͙^{M򦕷vQ/Wg5͛ŷWi_WIq}AJ|1cG#ǯ_S6>萸FZTB\|4/u/nr;/jT1tT=*菼?Xz&`ڙŰFfq.#z~1s-6x.oXI[0f~zLt; `yO|f t0F#jL k-Jv` bCC{1ap:bF{Comyy-< s1 3+i+y.tņpBr9+?Е%=.ܭgB>'O9̟u;ޣ$Ycv7nY4$2ze $nQ?-To4P?<%hAg4yҜǮڼ2ڗk1D)gE#dBV[Or!}u _ciK**.3ޕdo)>Z0hZ!^\a.7cZ2XC0w~,`.1>$w0$۫[4cO[i=zn䅅]X$E֯bO{5EU'Tw~h nlA~I٠ojb>__Ū_˅ӓ}!샢֝bY.~Э@.щg"Ef=]u}iy;8ͥ-|!4GtaB~6J|.O|{Jwuy!nN`Zøyk"Y|˒\e[^U0Pxn=ͫnpZ"%UH8&)UU8 ՟ F7UTY J~\ֵv t[]$îjx #ȸn꩟;ڡ)ְ96^{_&wRLL3Ouxs?֣yvo@ϕˋ?4ю+f$v{<2{a-c"ZG]̳RH sQSE 2VG fb⾰9Vi=,7>=LjcwYFߥPXS1ad6Y”M7tzt8zZ[Batf:6ͪ y,1_-?9en60k^ }EkWVk›m[֓dj״G?Ym:S՛[Mն._mY̦3;]~;_VWwٴj8C5%H{-e}~ֵNc< 7a;XUm&|l_IG`Py/B݃> x_mš:Ah "=-H.|:B+>2پ}9z93B()RTOKKYj)U49km(yQm(gv>cgyӛPu%( gdᷲx!Hv_qqf2kpXL HM?{0?o's 7k"mR#-줛yySqHKwt,b#[!BbQ|]$z}Wu:]<A!~Wq{ ١ L+7s!|Dv,J%,RmL^BpO4ߒS񊅾 ,ڋw!FiRI#E;1+#ZyK 00 x%l4dGw9{ćv«D_zl }u_ؐg TQ'E x]*.]w<#XwŜT튝suOvɾC췦K擖F9P!zeTG%Q(c9]JmPpQ+{]C6C=mNab2A]u43Nv~FFlm腰W$v9jU[?Wvss#Tc3_,3^oĆIwEl_b{glБ+^'0saiv+9xײ0 5@O%?|wm2𯕂nT#W͡BEkM~]@l4~1R&TH%bPP:~ E~M^"7V&"%KOVwiHU*4D2`uN}[!c #)dkK7[e]hxQD`9%,*L,1 f8\2EA>m"+wq9$$| _0ό!<& z]D($?TM[cxD#+r5HLŧTKa|u 5,,>W|[CPH1cM3NHE:%yZK1 ct^S%TyO 6PwEqzg\|CX|4qѥ&] VGmcuy_2wX`uL} VGcz݌ˍaCGCL p<1n߽J%87X!VyO EF ǻ'm{"I5cu_Pȩ6^;))q&51Ғq &:"hGSj繦a}7G:ym/D"Ed WdQV-{\D^+¡@&Z(#3Rr 4LR&hNG0-[bʵ#+Wg~*͞~Vq"MHKX75#d+O7_/yO14ғnf̫U$Uh$ 0u*C* 8+m q=mpGi*L'.΃ѱ @J /yöY0% lLz _&`iC jz%e EX$O!ʷm݀6@L' `c`H8xӆNwQIncNhM򂇷oPMRL}͐Bu3^bOAJc/4SφGyx3ښDw V\~j*vya#zjԛ׌&;5fTaymCyHFL=0 /0#/|c)a! ) q_ H)ҴɧyMb{NzBzS:'N/)*؅Vh;lh3f͘l/mIwɃs+4بk #`nƋLAf_9>5ToFE܍3?6rM{sto]%7EO_0 w'dRxRPʼnUz>4z_2 u4<P`SXv:Zn!~do)iyBha,MIdY:"j ]4K#StCs zͅSOM}f30"4/p',6hC֖LK) "bZp&QQ";2 nO`VSVTPlW #':׹,Ke Eܵ,oƋk;slO=ƌhP^Y~Z]yFJd.E`**!<^U`B4~.>V ^JF?uڊCm q8lKOrc` @F0Q_zG:|:L6E#gqP)2nLDZ F3cRe&2~R@P\>(t1EۗeQ=+Ԧ(~;4k"F"dJ!N fpwj$ѧ;ou8YXKTQx,G܉J+tOLf" 'm܉2:"hGS߰NX@01 >#0d=c;NL'B'(B5ތq;2F4gş RSoLL}"xm6@5Y_%Rƨ"`KuA  j~˨d 0⦎f O")r; O0&G(%#Lifi18>IJfl^'A N>f SL̹̎9 _R1>Bj uh"E t)㪠W|R;HtdWDhgWDvȴG[^Ho$=:V3[ltUIaRNHi*RҚi _Jks+p+á%ky_Y;>"(*\$yK(a 0 "XSHuĘdyOZFuwNƨVw8rû|5k[Zͻ~L?FȈq] vWR6 @\Br%ucgKCWbOv|:bԕG|8 ;;9*T|#׶ƄJPϏ#|z6bf~6n$Yʇ[$68P{G$\~ xc梗#DQyX`V,p# /lVې$:T1p]IXQ(sU8Lv$PŃ5EisތĴZ/X @&ssWׯW5ׯo/\e֯Onԯ~p.?Fy“c:zT`(ndX/R;rlW_k|tx:{9s7[сL痋 eN)7IR)&_޲֛RHG%}%}y_`xVAE)2 $p[`P|Vӫ7j5UYt_3ɯ~ebN>uџ O ?[VF-3jn4WCb~2K$'>enMͽM۹ʬ֋*wf]+MNu`Int6& ,7ET. ok襹l~L:Fr;oN:uLD$MK"¿c⤣f "+=bWIUGmeG;lG| y4NQ:q"+xbWQ2߹P3< ÿ(y^Vg>٭bkꝵ&C`LȤϻ޳`L:̛:"hGSjȜj2Ӛ߯nV -3"?2ӆ&'98"ZLvhZ[A$KOVwiI&8r =KxYƒ;qn[Χ$-O(^Q=iH&6tqf@~ͩy_zh5c^:tS_DX{uā\7 X1I4ozKVN}B;>PLT➤cr h,C|tNY{9,2˟W ~֔r@ZGWYp?#HW2 Z*Q0j 3H<0%c˔Ƌ[WP؟mo3,۝x~hQp,߈ZHkKPp8r1UͅGsZq<k-"*(znlj;I֎71 l3sr LonShkm F60dkRF.a*8w18@[([m3Yڀe{x,r5b8K~fSzxT,_>'>j&rὪHR٤BU$W3A+WOv<f6b"QzDO &T SYGlaF드#'&$Lfy拟!Pk%qi9ygQꇄ88F17E?Հw!Q6\ V}$ÎKI'QNWRF׉pNtF N֌qoNx^LE~ȭk5ofA }xSF~2$Q'^ Ʒl1}L Ʊv 3ã6F:hs)1W2ܸi1{xkzKZҜ9& 9V5SɨP])]^"Uܔ6ZԔРE 'X%݀w1n[([ŬXQU|pjW_y1Tq$Ol/3ylckCFdyu˭JSRU?ݝ-tkDqy[!}3Β^Հ LܢG 0VF,$]/ TWgSlL+lm381+}#W ['دro.v&qL#ѯ9UX6܏1Q⛏O?_~<>N|!ŗ~K5XÊ(<$(^9BDHdxkK07=wܪ/rTBk.#V$!24Hݔ5#m'!Y0|5q/*ǐ*r%{1I9Z6#\ebFj_L̯[bƑ*f5Q3zx;J22i|L8oh>Z*u0psFNaXƄ."}5GGܻ@?]T)DtG ñ*YMjd hf݈+/]-X>_J8fM+6BғYrޜ8r5چVQI/z 8/krUc@$fϓLD4i#qx˴ER#k^m 3rBV-R2Cy@lã`fx_܀. (9Pj)S~ T^e:ϒXr@OUp"O$C,W=kP䷊Ҝ:-ɐƕ޲*9䘹CLQ;V뀏Y>,-lhglJ lrd%#V10WkYϟQuRՖ$5Z6+sZ3tmmV&n*\a9zx㎲jLze笴r\|u/yQk)F .m/߃K/E35`fMB7Y/ VyO|hԨ{@X@"U#6 +wj/<@N&]+-[mTU*X)bvYxzRNA(8%,kƧpn(ܫc` X1b5hG5%7516ҚqjSuaa~nB*{?8[d0\sgxQN%N{'`Ȣ<#zRU5|p/dK xupϚCMC~t qf; 4d@bi,FKNFv<e_˧mR`l=Y$oi#fnD1[}kҹ3LoҔ@06;5?['^sL _]P6>Ejvc @tEt^Mź"vt>4Y嚟Yݤu tU?ÿ(r5ԣtUߦ)hi\MKs>ndXq) \=cZ]atV=qgpn&-134<+WIdmSV[O@JZ 3Q6U<۱ ܡfGL=՟Kx$T wkn犹q(tύr1}R3~kjɌݎ7w18 9LNxFBVԁ⡒ǷOwpV;VoRr2^&(/a+-XC|0ɁX|ԉ)h$(I*`'ǥWG_V .-uL]L7z/a~ v(ǯl[;H۞o|wk:s7Fi[8ۣ';CKHRbW_唦imwʼ!H J(KNAw<0=4t|lDJM)_\o Gۛsm)jipÿ.l@5䢇Ǹ,M|]̗PPGշ.# bxN7O(Y|NƆQI3LF(?`<3]sHt!H~x,KjYjDm }xȢG)ͪ!8NX|\xY(!U/B ΰ0\}o_QF{_F3$qUF UwUvckoAY8}Sy]kc}+,"Pd W`_# ڏM7|AO̠oW.84n?XͲ#KuVPrm# EO_CU#slTmj /P} R1pscm#xhMO;5ŷ9+C?l-ÿ.RcE8?{mֿ?ey_|&o~no!V2nk*RC)}LNګn 1*D{yGkTQNDnf*W2xV&XO[Q2R4w(6Q>FoCd=~1EnΩ U埌d cν];b8">v#˪8kj1ӉܫSD9O qOpCɔv(1tZ b) N2kr) ?:jԅY.q̳kKdsSm>^["%6Oh0".TaN\' x C~v""e'+%$x nHf.dS >'"8i+ΝOV_zC;0vp=Ṃs(7j33sTK)T֜UT[Af7O0e{T-,NB:H Ccx; s0U6p8mfcxC{]f+8%V)\4aֻgx桱q ~ԏCl~6[e7o!k'Տ_Ξq=LhL͍pr9_| Pvpcg, 7%1F!FoqkxP35$z 4kڤ5iuCyJJ e.f%^U$)%I2fV'{r|O}{5"Ո!6r͕6nϗ9S=P+m+mW:Kؽ{\VV jXfcC|XGGΖi QyVD \73$Ps- qo OL;X(xΝ?]S331ٞ]9EA]J^-HXYll}hcHB%T+t,dȾ7et{h #klqKlj6c^b>T#~JǝjBz7Pf/>B霉qW|Jkx^ ".sl!D 9;qC`-q`¹NKȎA9ͅw; $೗V5L*2 G8*fc4C\d&U&Qì{P<:Г)ׁ+cMTҖ*jF~He㥓xo"O q׵;0 ͎tձ(@PN?Neq1҆Pgwo6$͕ԆD'V5''{hcvw'p-G^8vׯhn#vo>uը3o p.c%)iIx]Iyň5s+yKq/fPe*.+q݉W\ڌI!|Gŕ%_Y!pe>5’8 o9ÉWF#lBd|GSAu(^O?3njT"ߘe{ Y>(-M 儝 /?sSfwNp-:g {$_Gd,cNH*bf$,4p,U*1J9VϏ֖z_Iwc#[m{W8nj>%.uR})_|:_$\$84'qf_\iC+QRk$:&/Fw?cxKtLVf5a\juk 1k=z3a~}To& Ra$lF 8 b-)a[^>/XHTEͤtY(L_L]VaZM`LNgC\=|; Ob|h&3ybrX8nlY\䩚Φվ+-ǃh+ǃ!ޥ>UR2a(K:)=+ 2KB֋lB4};_p,'tϙ(I 8\=o'0`&b2Ui9Xt5 x4(L~Z==zz|\,N>u>ytIUg|;Q,ѽzcܾhY)o#`XY"lYc.QI;P_ga^#O Z6:kӦOڄ85sY H,Mbda2π֌_ u[}cgd#(U:BLjK|gBA7 WmG S=<"- U 2ej '4ËXRPPΑ 50W 1 ! $ he2+JP^rK!ƮcPw6ɓERGB:#]dJW|@sT;s,٦YBͥHD urfXq4JS%ʡVc*]0E}@)*T8\>Cr~jAQM G`I,(sQM }"ݽxo)¤׮XTǫt1ZIsEVVy5#n6[|[nxw[Rp(2|WاÃ9lk26.<32BALeyR2!?66puA|qZۄq(_C 8<NjN*KB:q¥â gT0*8y|iMܮcJREn,qIkSI!XAc9uĻxuf^!>+R4M/_wӋ!FϼZQfU,fԶe\y)2(B2jUn[e"+&ޥA)R5+UJU|Et)*mTg_[m_O߯wem.9hЦc9dʒ Lv4c3J^c$"vnȈq#nuf9_4̫}(פ\ΨN#Aˑi5>J)<14$>5W Z0SJXS{h'`3Ucv(νxLPKXo N >#̵= jZdC]O2' 6Y9T74:a>ʺMd}43oa5_1usib7~p{lyƅPC_;Y?WgV}xݲq[~͖ )L=1B~O*n& w:)΄:Ny%-)ɲOl, 'zNues!4YAa8ed6R,,9a67st%60;'I|3cXi PGyoiJWޞ,+O6ZS:m=&'Ge#ZusZu+^|ʪ[-R*b-W8+VT`% <[hrI-Rļ`p 50MД3me; ;u1{_}Ͷ]E\}'}T7=Paɫ6l-L*iU™`iAv.sAq-b˞|:C6/[+E'1ư' AtIWsDψTԏPCy(GA B:ci*#$Ff}/.͉UUۛ2y{I26Uf.sDQpd 9Åt}mEwL]'+T*`V\"}g#Iix'W'+$4zlTi 3W&'cXAʝ8~Ocwn}~}4{eMya*u@MŊOjbR<6Tnyzϧ1MFi @ׂ^OPKZ/v93?{V(_"Yjʈ;;GO0a.^TF4SPoVK&Pn*_"ǃMuWgXgR ?z $[pOM-xϋ2!C%T_ǂ"7ΕLl;.Dž;wYPݩFx^I 1Js… R17_{~~h;\dG47pOd"ב]D{tsZ骐N‘].7 |1n/ )¯n="=2pϦ9]N߇5HkGPLwN:VkɅG1v"T;91G 1pX=@a쇌s[x;Usa w;}rOYμl6~wӬzZ…rdݰaL X:&s("M ķ0oH09픏>%;]kًKxl i 㵃CLA50㖓 O3~JmnFeJ_nSȃw@U8{/;UwsaĈ"r$N_cHIcQccH`(·$)txhtZ%f<> _l=hϐύhJeۗymjLDUJJF8 0$W;j)zdd+~a%xblnQ(&FO쿢63|s<Q Z9Om5˜Ee)'ru!2 z !::C%Od ME!)u@5;_t v?Hݧ,9㹖S[<㉚ =22#Yk{=ɧȕ}%E%g%r*۰9*VIp^lYcHѻX l`! #9!d1B>U7WN7=22tun\GvcȈOikKTRό,/%9!TЩFQ)H@1vWrj0ܜv]܅RZ,%"Η@?ƕ`몊eF5v-6  aU++%MesDIc"\ju FΡևό ' Bm ` bdr\9F{mg#}.#w} 'wǎwlC (R1$9gƃɪ(|MJd`bC 9*F)&%ixDi63 lSO*BDxս3#;S[yk葑x{O1U jy5ĂFN֕GF^pto֞^ :#@F9Z XA*kLlCIp|>i^qO5 [#V2aW߮cJA΢^ It0ń&ΎOOJzt##38b(49{ >~ d9΋j .Kv4JxnvpcÑvl&Umad:i"aӣu B@_; b Ξ5s?K ~G[<gGq\Z!LdUsdl3Z7/ݻ n7d::rmj?]nyGdF DANĈV3@ `<(ey, YɱZ\ǡ8AޢȠwxk73,E{60dqCl!V!faQą~μ+ό I6XʪtLpFq`P ԪRB8  QcG.h6j9 .[,>tXʗ ]xE42ˆQl m5 Rff&>tGffH7}| =Hx;F3FwR$; "}L[yqHp%)һܯrgLʥF@+2lD@l F'o722ep_i8/ )DE)JKeoBn}zs"鑑m¾Ռb "Ud}#n 5Rט,$yQmEJ}ȼRQ|c4AMP5`w|UU'JBBZW@ܬD#]81hQ 'zdd'MփeV|=oFJfHɱjR i GF<ҬƘh##384Ij>L~` m##38 v-^g"t@q]GFfpqQeFZE}se&ӐE4pq'tGL~>!! QNny+,NAyQP J2v+`ݼ8hR'&dLNsA9[u,|VyGFfpjsLNɪ9| p8u̐Amz e6mf|#JObjf6Hu]ނJ̌NyK. 4H(` `mNsragŶ+v##'8,qzò =226 +*^)9g!a#6aQ"m"}֪gQǟp^]-2ꑑu4%Qӣ)%NPax 5q;]r σSUbRU6EZ0Ž.]͘)Ei4Ҳ` Qa_!dT#1zFdzu^_`XQa=22(_hnx@8(H)TkR+R٧(~iS7p;oT;/0S2P+V<##/8,ŕ׮a٨a.dHb\KeL+@*a-8q_Ff|HҘfվOYRjf 2+'/QtzddoQ.E{)r ?l bp޲N+IQ9v 9P[iZQ|iKJ[ ";fAgɉ_6~9'EE'5rq#*3Z Km~H.whg`- 01oh{dd'[fՀ"*C*d<~f\L>ieWD HP|)fE ┿" NƬpSWlv##38 񵀣L*3Bqzx|!RD6}Lcuߞ G9% NcDT9i(9C X+>`k! gɴʴ013sLRK7 $[(=u7ڞ\u Lw^y23W߶_]aͬ??b/mˬy7n);svTfB\y &th`3 sd>{QnN:}{&ofY5b+oדeR`C {K3[mj>s -,mFc&3,`B֦_'`V_ZO64}v>I0Rqx7\ުzvmף?|{afГez [# Ry+Mc6S^ÀEN]o {?@9WoT) }SQb\ 5wr2GpС$6KM7&QGpnK-IE„[Z3k ߇aiKwe_A'ūvdS#ljҴ2w^OLw=;43VF`%<{o*P)GQLW,pj/KJ9 Up:(%j ouehN(VR0gǾ y L Ouu;m}T'hٹJu>cso).)70ջ|(?oN6ss:9rf@DB*kw%lY1 z=W'/Lc\[p㪶s&J.kfk/Uc8X{!*_O$@~x!7^n 2 b0w' S5Ii CHo8 D!>O f\|= FRWՁ[F*6|f14luXx|5:upsFԳx'~}:]3u(UH(`lp"1zdAdϼ=Go[X]=뗸(l+w G[b/ṪhE(0!l e0uvX]; T+#fhT?;.^XN\L0gw٢'-O8Td XcRmhǿ{ Ϳ+I~8ڋlRwGV@]}6雩obK[knjخW@|w'[NT[6FGr Cqb^Ü8K_7k1ɻ{}}!ԅН?E 8)U؝~q~mM"s FRlζVym$v2ܕ*ޚȘKd"S!IB0i!ǯ|!xOX]tv{#wZMq>^w=~vc[U'! [0*T0]w#aj#T炉 3T=_LQ#OSܑcej?tn]!Բv\(,kALB-D*cc8"'ݧwF-vADZfn" K0[֗7_:_nH'9)'=69fsKtY+K,+\;N/2+}`bH`$! <ߏޣ,o(J|"ɹ,v?I.&, ^ǃBJෙEe;/ -/)= <~ yrB`'?(1WAݧPj Z_N!o@aV>L$A]dʃ-; /\)Nq簒7Ԉ/) 6CNջ m:,Q7~bNt/?xi} .s[פ{mPX'ux*{-h'#hՓEZq+PXpdUPwԷ{+Vޓ_~>JF<.%<qs5L]yôd R̕`WscɲOФ ւ=sU+,yYn_՗cH(ɘ.~2*hKc7WANn1WK]Rs>UXiN wN9̝ s2D9-2l1F *"jX_ђRQɨubXUZj-XԼ 2 )4\?{Ƒ00˩e؇ fOE]DdR+JvE~IQĖhNl5ݧNUթ2T&J6zj q!Œ\jB'R,~Q T vJbo%(nn`gݧb8`?8z6\q|oo/r;oѩD0=ݦM\}w2C?n{u~/Z|k-c  >߼秡{0n0|ǐ˷~6smZ^a^ xcg'oD?m߽{TvKy_gi᝱ςֱIxX5&uu>~WuVeo }u&*cϴ1R!ĬCE4cg遚%2Y, ~*gMm3[qp֩vfKZe#,뻤vw|1_ej>?۝9#$sS=9_f~7u7yV85qQoCCwqX?ab~EvdmxSC֖p6`͵Yݼ>nXkܹx]f4T"ణR}hsml/qm7<|4oqlC{sG{_xuy}|}DqS0^;$ C1y(2kKH㟶~|$ݣ-{zaHx+ިk7,3,sfkz\[.˷{qٷ[!"~-<;t"buvz*y:P>Dpy}g)Vyp[7ۜ#7ŜC6 >ԫw@hR@ҁuؠ̄)^3%H~>F<5;pȗ㽃x{{ iEӷ r WfA B2ʦSuB(VKܼ-$JU1Ա fwpzށy*4x].v hi}e5!glruЦu}Ma=IǀbK2׬M.8kU$NP֤ 5:t+kMZWCB'Fg˼p~1oHFm#:,PϋpT赗BY*w,˼XZ_2/Ҙ돸3=YyVuFF^?t>EbJy|mMʗ/u3s>2򽝟~Hv0=&p(-\_},"v7𒎈̱$̱F ci edHHX -؋9A'uNU2m.r{s43O(޳\kiǵ\k? T*052ӗBCJS;:Mi_)Zr `J:jR&,\B^ 2VDJP-YI"_ n PPT]wrbTT7W_Tx_korȑ#UܴdB;<:k`~'\7ۓDl07* 6ZOǍ9QnOrÆ8C6"_\o 42VWk! ܴ5^+s~KUsjoZ_?v5ދV=w4VyϓukN;MO˷=`Sfh<) >~+㊹; i8ϳ@cGCƟϬs|( 1liU 5*A|z]JҦ@VTeVfkQ,&|%Z6k># 0'H Nsk(q)Z:&'R`ym)ddF4A ogߗJ1UP!"ܦ hL\>5n,S4 ^i)p';X+,Tyᓗւ^ҩdtFDU*4פnEƖb1d]RvTZ6hXC˷oLHFl5crIJ[׌R8* щCbii{,[zJb ؔ L)J#I>bg Dx¢%%2dY ZH[ ӜvFhUXPQɔ Fi"8} 2:ּk94Yٚ,J<2`a"^#xu@Kq@82(}tE^J.G[-f00 3)<"Y7.q _4 va-Ԓ"4D iyU>8M>)Ù. K}4m #"ArA}IE@V/VgV<om U\zLX:7)@JTU UCVɒ3L0)qaU0~ v?noV8?;o ,!VX(;Kb f R;IyH!p9X.p /o%LD2Gy(%'&b%scY x %@@!Mv% K |0bF z,n0ZZ"΃ ^;=(,k7.,L$fJ$)ds ` @Ao,Y 8\-J>",0l" sU&B@:Cє КKd/t}DX#jꬌip`Dh3ˑlJ& OQRp hoR[ ABS5y.ג&xbȔZ 6mM:ty5_f4j\m̵Lc3*nnf= kp0SQ` ~(xv9Y:Zm*T-G(ZthvfcL y{o9)-P1p78%<%2` rhSbH|(GЍ>f$IY.j:#S :( ЁrHAO!HHY@oڬdp4 *O1YWT!& <*@jGAʟt׻ +/V0 .)&PiQRSdLY8z/ϤcPP ȣtah#ʅЏAT80v*`c,XJ4xwE4*Ei H>{*ϙT'#)ɐ+&Ts'kuNkjyc ڊIC`mqa :"<02ja/ZC0)"`0 ' ZS{(0qK^J nZ"i0.3;SֈU0\-2caVM7$C(:)@q%Gx DBFqw#LϮhXo"B [!#f/~h۵iکYfdaz˅4xvӷ?}r& |F`hH ^zCam6/te=?/FoFZ|: nGB7b)}H돸Қ}WCwE"]ȽNz':^'{tuҽNI:^'{tuҽNI:^'{tuҽNI:^'{tuҽNI:^'{tuҽNI:^'{tu_o4U4I$1M45拯N:?b&'.? y^mKТdY6 w]/ԭk\qH[YbTQ OI?* fۼU62|"}h~+wof >Y+Q앚[V3U:ǴJAt*JAt*JAt*JAt*JAt*JAt*JAt*JAt*JAt*JAt*JAt*Jӭtb*8Wmar*h }`U:"+]VΧ'yU{Y*Hd7~S*(Hf㤎@0H@em%F[k;W}(P$ޱ,K2eӥ0%!;Q!wֈZo{ߪqOɕĘ.~ +KwmxFX>sO'u\+ Y_ tg.HO~]^>k ;ϯ+ڞos'h23nt~ 9(7(?Jt Wcowyj)!73ځ^7qKxx, %BD~ΧBK\J'U"U*ڇXk-ܽ-jՍo''GWhnl["-*2D7Y [L92߃A|jҩwj:ap."-R ZF'A$i4oAj2<hP-n\g@8vWKEk|&qςQIGc.CE⻣_jrpj^[bl?=RBl ^Mǔ[O𠂝8ˌYF x`DTd:݃ !a~X}-ʳ,uCt{H}gKw0{8)'gWN_>߾oдC~WG ,?ɶh#PGo(|t{D;y+g1.wȝ:h܅9O w.b0w)h"(Iܫj|z>($<#]I0:CyOqw9JpwBŵ5%Q[;Wٛ5Mm=9lrҠidײ=|=7Ï@WywǬMdgolMԙtzF?w뿬 PY)GMB] Yav P1Šҧ~ |<`Օ-׏*ϵz%hm::-276 I 4 nYcJʞht&ܯsOu8e!=ye7ܞ&CmulaXoP|)n(4Cn )̈́`zȄ8`?dBB&J*;G42C&48Rk\psUԶ/\Y,{dpZ~qRfDsx{~6>NDh죹0}>b~ĒEŋ8Fh:ɏ!@w/ˏ~{. ]s&|d|IXjU놢^g;\|"MCSTv(uxwPSCʼnbu5:oO9h:;=z(gC"5:QqyĶ )X7.&0.XVb|*!X >(,Jj֢'RROl`s 5g`wp8PZjՆ.UN.(P^_Bֲ2ְPB{FfQ.PPAGiEіx' [y$/MO!swcJϙiCgPVs77Uj"J*EFw"x#6xF8P"౉6g1p$oD sؘ PmL5<"$:cH~`y2,H"D[,qoz 5;*ԆL'(pՁ;s`DaW+w" xCk[{ 3K)j8`Hy x*7T+!'a2Ŝ̡V$Qخ҅ y?@B %rD dL=EKJ J5(=HsU;BQ' ErNgN&J\ nC$Tm핗BG!\7i +>BЦ!jܘa47`P"2ā&&g ţ+= 5UKIدOଁ,9O`- r??DB Fe<|$48j_$ZHI,e UQcWb%g(+ت3Rl ~c *b $!E bVk̈́QP*h$0R 8ʘ($snC$3Y{k̳24ºI CSA^Bjm!iQJQw=~~*A-y@2[^LBXGo"t&lѐñc$ Eײt8\k}w-A]uUHck ȶMNT.Er䞳!jWO&H=e H(ugBP@f@_qls5FyO<PTUo4X:l((s0DBZ~gcIu/3o>K?@B N90| +46xb$ƒ= ; j0=Ѵ1aV Y.O={~~3y DД1Q:/{P w ?3YP<Ϻ`GY!(-j~A,Fw}H^p][ 3%z"&久,&DӓPds =XW&){^" ͺQ)s sIJ ="x|[6p{ )!Dk`{1DB 6ą*HMVcb,$WuHH]}kE IB|P_aT% mOV7DBʼnꀏT5-4%JP <0 rHqaZнp чZK(l'ѷPQ\}̀53rIG.|sQԏ`se3\++90xas:=}wFϟqIΟOa1v"ψ3z-)q۰?jƴAVNrĎ18tՓ}y`|}؅*KY;XHz[UƻKIjI~Z7ѣro4u%ĺ|q?x1dDzYtoxy;z ~1wb_k"{u?~탸>%Uml.7'u}wkIq6&ӣ,Vp Nq҂21|Ll.sHj.ׅ8`v?`=8+癤|q8BPs2 kTdu2ϊ@^Y %Mm)XNNfR>pu0h?ppq]Ŏǖ^7HV|thPkL4W;k\}sv2N\zqxTgx-YTb׭ &'7`||k2'Wؖ7?BlR-sΈsrv28uqb7ou[,ԙ<['M].آ|Oha,V>?9ӷpWJtAAهx`PE)+GA[@LgC19s ) q0C8 ;h^QwtٲB9/W)D]-Wp3GI?;Ɗ>AҠ-H!HR.]tjg\F/~'Z.I:{4olUNƴ| ×~!ߣn:اjWLY~?Lo1aߖd]mo㶖+*Ewڋ[٢3m]DQ;d2E%ʖvl:<yC智Gh~ѥ֏ zDQ=f9O m9۟_N l~OeKR26p<5R +|̌)mygXtxi$ ¨Da(,S1 T2(d)Mc͐PHʕ#ۍ7ANRTSfr'I#QN(FF(JdfTB-ǛZgs36Eh85bg8#_0Q>}m1oN^A1 -̪@ 0e>Df>'Zbcfl&EE J~2yLbqþ)Ag1%oHp2>SNjQ~s1Vԇ8rT 0UID9c'Dj˄$45(ʼn 6 R ǑƑ3(o߽~氁%c/ş:n~.}*ܮTHTORQ@vK9c Q*PH`HjR)q VA6𸟦X?3]WϽcr QmgfQ4ܒMx|Sc02 H)9R2BB@ؖ*EqͿ3pj oM85g=,i ],KUvfNP v󫊧ݎ߬MxbR|lf\snxRyR 2Kqrw3;0hLoZ QHOAQB?c@L',\DO|hpohpu<";;;fjj>M-q)3L-\3R_H׌6|ODG -2Z/D[1#\ VM =Bpjsxy =;-_7\6aoF>FӞ>*;[ϴĴ0B(%ћ\5}IeSr)E)OzRr!A)h2BMv2.rvڵa7BŸ=T{.%lf3$& w^ui+o~o՛p/N;DUN冿ύ;M?~o_W逊i}fO Bs6rÌ:R'A1O6޸s+.}qVZvJYmq5{< =IsWYWz\lKsDI엻 7׉4MAA;#zB#"m?ꭸ7ꭴL:R +$ hKln#uc;0r,v$2&k5ncDgCEQzY'㴙hǮm7!N%FY1 Knaa.14xtT*vRd"!#7OVkQQ$Aiw# $iUd[+]y K5kz߅qQ45N&Ta|}mʎ2SH9'w9f638t>7s2 )id$1LcHb^ڍqbdB[=ej6 h8fM :|>}ä4M$F@@[Sű̳uqXZ-?&Fc 䗜EʌDʬYi9:HRH٫q>uǖegt5.xmwQ^@`)n#1A݈⺖ AG})M.\\ Gø܂:|\T7 1KoF`7+R1[w{ 4?,JC{ + W祷u(x0VDsH,7J}O{^L IǨ:yW ٻ? J$6:]E]T+.I1!~8 wuWD J)3`qi9OuŮ>6e g_FS`9έWյע^ *zp2uL42Z8qt%ܶ J ok[ڬOE;(kVHE;eo :Me[͵jfPt#Ya:TL7SIe3VaoCAi`kBt8|Uu#Wȧ ұaAA@Aҳ Jw⳯ mR7C>KJ՞XBԲZ> lxS<8!rhL٪Pn$EC  &k8V3vՓ7KrWW nH] ԶY3>ޠKսKՔϝ^8Cv9-cw"_hK[A`0rL;T/yU] 7K]RS^Kg|VSZɾNX yGʗԩ]PW_?g+%XUlH2NE tep@;k޼äP/pP[9j6FpgZ|隍JGH7n PAmZ _|])t-GQUhiG@'G `щ^R'A-JN0/WGtuk)n_uY.᫴qz&A3gDk} FQvbg h>hU;"Fv=KksI'Xp쯽7_ЮVeS:nUrN_NpĨ5l5 $U!N_nWn{Gę}l[/]VkP$dFk{b}uEL 7JIc#iG@@'|Qz(0H>S:9V?M !)y%O_J`i1w?10? hē/mU4bmz?N.`6)6(`2kO0OqԍVlr?U:FQ$MFi~)Zp<*PO314]a:5=ʞj8|dwwo,zY0̂7tLNJ/>m2m&7o~ ,$5$֋o¾̊31/޽1(tu_ t`3uYMؑs\^ɌgQT~0+8mKfȄӜH#@JNI@t䐡5kzegqlg&xPw>o݃⑷f'}|fhN \~?~}ij~q?CN]fR—1lnM1*VeaM1s o̞͠AM?1lOmwGcw "q!8Yg CfWg|;k|wV]{VoB~)8z;XcSFfu"h"qk D 4 q5xQjoH4c+p )DsRdW%z&~;03'2{T FYXqbh83]b7z}2KzYY+b*߃wvfFʻQ/> NnIla+I&ʬNUz^g|T93!PXƉd!SDʼna "c[JTJK{! Qged|9? rkNldx_~j/BΕ¹[8Wv 窼3o@MP AST C S$5I,ID@$lo> ˌ|qxM_& |c f7&&_~2VґSYqKVZN=ga>zRB\%z e^ ,qGi|3Y:^ޞ+爝;ʆJ,f~V!t[苨Ǖz\٨Ǖz\-t8PPi IIOT*3jfG,Աd,J#![M_h Pr! Xq)KJ٩N+ @"GVCBBƯXVR8EuciǬu}E@ (Վ(\2oQ~UͿ /*-{7Aw\6b0I`X#]V=|e,>b僩%rj!۫|E? H0E~bWӡÙv2 ^VB/`F*_> n|v L %jqi)f\7B|\f+-~?gW7ACt boϐg̊z}f%'}fd~UlA6= s:qeĕWN\D,b8"FRc TDy,X PTBSH&)b?{Ƒ`G[ŀۻX "HZ?G"=")i33TwU]HX ;B5jW^cop=A5ps^^#׾:W8Pf+uW%GIqU|3opva1QJ8oE1&sMO[ED2O񖠡B4Q8FKKJicVST(;Ohd P*ur.k #I \zTe٦6Y N-PVGHõ9i $cK!v)Gp4dٷL/Y6f8Y.UZ \"WiMA*ּۘY*e F)Jl?8+\FߥwXR2$.|Gz ]9 ws&FӽĚx8>&Ym7ⵏ4fʻZkBkцzI6"5LAc I &.I* 8S% SΓbt:DdOo.IJJpfq#\Z.wLFz׻^"R]4J>E9^6PoIB@q\F4@%qɣ]|r tlDKJ^K(A)`$[&h\0bfsz'B*ϖa6ć/bRru;&of.7ע)+hP#QV!wwnx}~oDz\qS7oϥܓ [-Q%:y-Y #"fT{,%U4(DŽ TB,0)8Y(T#Q 4zhSƄ$PE̢(QWE j cSP(T+"8GuQ*TrRGzoGaZ+Ej4IAg{(l/ÕV!f+ltR'}Z;iG,`DKO  qD7lU@*[J؝KPJ^'Զ=& P0Y^qáS5?k}(4XZE*:T`NE]5oWt?z]T 0pm2YFC?O7Vr5@Tv̓Nk@5y7WW^IL"*Acl 5^SWc\TÚxm䈅K|bQļ(Yz~LPP%Y%8 Q *5 G}ZVRH!bv{;#߽]JpmܸBkoϞ] tvs7|oV#h#[t}K/}qȋ'⭝e_'؅;4^^\[*syCM >F* Asb0mnѻ 0q}^*l#OBafe6ϫ aI!&4=I5ڨI 5]YȾۃZ6F56 E|n|vv[j69j#۩j펏x"~vC)_ѯIQU0*t+x mv)En* JPJ<'_Wgav%ήRʧ8XAn:ﳚ4߃byr|oͶ*;6,]7j/B'=R0ZFX IE71kS*`rAm]Bډyn[@_=gѨ {ʀ*& 6F2=LqOG x%ei7'd{0pRŬĺG/0JJvkN jjv@e!ZCFoel*1|18kĨ!N5h|297?~RN(ݱ/4LC𚿝$|Tgt?51$̷%hkygFyQn"W#}g22,Ux1Fam22 3}'7 3Jn&XBN ='h)`E\#\~ز#t䣢+Nݵ]2R}NV ]e/j֨LNtutŨ\DWh;]e&:A9/x1t Q ]eCDWHW Ao V2 Fbh=FIU"wTj`lb&'B| ˭s`&S x29II:Rb&9 R2HT ^9)d)7+ڊHDtD*lT"4|J\IphCN`r;C|٨Pv4%BI$+î^C7k*H1HcVx? < pHbE\*D%91{*7RRQ(u*pѺ^+QEHZ& 3FkTSdG lJ >>yQF4+B j{Ր}bM;OB T{9V{jP)"t ~ #`!9Pp%-hT%01' N>;DK)(T >lEAt,2\]UT2J1ۜ"])F %U,y1tb*)3zN 'IW\C#i~<:pQ2dy+ơuHGՅԀ KhZ??O3LI@Xr"3\QLD`F w-3J9YV'koU9Cw;.Rѥ44MZ3ؚU67ݝ'3"d5s_8+4JN'ʭ1Z&(KU F($m(4DVb'pǒ[=J)-zp9+zh%Q6R&ᄬE ToWv_-fguٗ`ϖaHAV~|kZ Є]~{q@ܾ> [f߂Qҳ)3B^?{ ;7k>܆-K!!M:)(#} }{.Ÿ*o;MޞgJ2\JQm9vՖQ6l'v:MQG[S()a\\R_IPĈ*:+V薊*&NY\_r9[n[Tղ#8O-st}XI^2t zK]#ߝ!2- {{YYqVWW$ ;1|ߦJ3֓6ۮ4A[S^4J@ ҜWGMm3#ke7o}7v=)P6;uQлvຝM8oO]ԍ goo^o_l+S2ݱWi.iWBq)KY-n7`V諳0::aϽtmԩٸX|gWMj$R?{Ƒ MH}C6vl0jQDd;jH#^pZ @"_>t„P^&/+6b˞ur&^y%Z۞$>O\G;i!%|)M& / =^Lo%ԅ.U5AMatɺ7j_h^FU >-y[pI}Kh.ߕY^AgԫO^LV_t$ md>$(.ڍ T/q#h%K؍=l<- Xa@i"1K-@%{~կcH ]!J{:C0wU Z@)hOWHWDZ%n0m9Ium}-خ3Y\2Y(!e%R2dq-X%V*qZ0:% W2SSXz3m$G`õ*xZZLJh(/V0S\N٫"ZNWpӕt%K^^>Z%L&\}JhyևJֱ|+YdOWǪ!Bt=]!]bJ40=e5ʮd·,xDWX o CWخ+UhV۠ @S gJ˭48։M ,c<0QhxTsU&JTӤaBTF')lveSn(Ni!O$$2aE ·P hAsQp5,w`9K^Pభ^kSQcpdYd7>ݎWn޽T猾Auɓ 'Oͨ$)9' ƿz70mbƟYCо~qO>O-*DFhc^ cJOsRV J:aBn<J!UD*j~#l+}0nrGԅЖq%q\g !$ǓD{#`qG9:ݤ#])"-ğ{[ruEj?GBq]!WbդF'BWҪΐ?AW_ ZuΑ$הx`M(e\#~^veklwoZ'#+E T̩?_BD+iםBDpWDje`}z_VCNY X{UTOWǪ*څ J ]!SUfk#=]BWL AGt3+L1mt6NWӞΐ|T&nY\0TLUę 5eaI1(61HԈZ1T^(m6uoo3bvIv.]>])ζǓmk5V3mt<+AM~Zj0B/W[BPgV9vCeyG}n7RNAa \d>`K BJuQ^?k4>U0gB\BWVT^Е.z}A,5h5QU%T GZ@WcUO'ko J ]!ZmNWtut)#f`oaa8lx#rExCWs5v\e"x{rm\URj +½+ˈ/thE緮RڞΑt/*Dn%u{'*ƽqe#F3$#ca\阥,UY*lJE$H.e QÇBe.*[@*Bl=@ x!=@Q1cfݰ\l^IQ՚8y&R?-%:3Z,ƭE"ڦZDM֞[ k1'J.f5t%1ꡧ3+Zs13ۂ{r`;92JFQ _]4yjx䲤pQ>Kqc2 060s2-2Żem6ܬRâ /A?N.|N˒WM&"v-M㟤Y ~8>.bP-78uy-|dRܛt6ʾl2XS@A|? 46ɳiW"KkhiD80L $$ip%P̋8R,SjJVp(=#A۝\#fbgB+[Taq sJך4)ĝmGE9 .767|.3\B4)2(| oa tѤz?xN.x~_/\n)7,iu$BHQ]oY\cej+dSW2Y!޴,`ꭻ׻҂Qz1j[}b5Ȇdsm=C!(\mנr;4ױVfMrbX\x!)Փac3RzN i̙lBC=զ"j~Ո,S{M%PPp@?-˰xT2xL/yEZu1>Dw3?k&MćIF 6Li0&0͒scSc"g4RThfBH6O]ղ=v+7/z5xG7B/UZxj0J_e՛ QD qS(fFKZXs2EG&Q&S8c\qNYah5a/gh]meh-6h2af30+|d 3{g8'+f8y&60,#Ɨ-\.We O9k5Vrv3=޻5P!.=R'LӇg:" H I"'$bXPX$x%@2TT'ǰd0ù9d5'8ϡcsBblփWzL'9+T_a5`$ \m^0~En*p<iC Ia'Xcey {7=~[)š)\ۜmISDԲқK0Ect~* &`wVMy_YDdUѬ$1&~dt7Z 37X1Me]%ijLp*E ]%LB_܄n"/ _7PVF)L:)Q] y&ElvkO1VEg}^ Ms:nqك48 4O] +Te=(lbI >,2!H! n:s g=;yXdg;Iek ,NBEJX'_jì H[A= `L" 51Ez3'Lpq'qmi~ﰢJMmE)ŤjQSyףu. ^ʸU)%/y/@[)2iϏcv> WpÛ)&dCJ+t05ן8IŠͶ!pEb fha˪x6a6DrJ^&ݟ>4Řȳ^H^~>ŷ4FJSQz~kA\)+dV~Ϫi|;oA)eRj=7OUi G(l_ CBlr`% Zq>ֱ(s" 5;Lm=n2D[Qv q^hY|Tks ai^]4A33H#>,"LfO^3(.g"Q08 DmOw= 1/uilT)\X[ꎻ;D(BIUkmݺNP O[:='EןAͱRcFW_@|=Tk26K_z%puv\,ߣ6)5 l_-6GRv(k.cux Ж9h]@Fǥt'<,Uy=yȎ(4D6z0H[XK9K~Xo\)٭;*eE5gkqgtT Ybh^n sTHh*]D#E>|M?JP͑{e\ ýt9ܨgC+@+Q%aToY }wY'a:PepZW@rjalehq Sxj[*ſxlA*Y F9Ar'\Kuu0\SVϥSJjpؕ5]!n \6w\3cq3-t`aUN;{#\1?#w a/l [G6Ə*J`NG#{{!Ac,pq4D8!ukZ r_J6SěE.-*" {vToy5;h_: (g 3[wpvL1ҥ;\!qYj^n򈟗ërrެ~5Liǟ&9s#15]6Ǹ$˶2Ɲc,_? H?~N~F{o({tnX^ `v3yQ=~#)f3A S(,:djʴ9GN{fT8OmqOo6]P, sW; ^]q1iC";ZQeOO:_tWIH{cVkN~lX^UjktrpQ{E,ż|uia暻qWA"Iph:c>H$iÇe3RʍSMJ XѭAEQ(H-x(w:~;IknC`!!Ê/=ʣ/:0m:`J3$<$0]pleΙ:<z\Y9=/pt9C1)2XnԁjQO_)j"Y$Byp eP00wTp0(OYנ)3TL_?)ЏreQ̎с"pͅfQl# 1 f>w]n}gg\MT~^/r8`9J\HUq&e`^NTTX+Ä1u| IEM㪙V:˖U0dRT(ts]v L>{z;TA;"c[!A$vTvoJ tNJ>FRvL~:mtN q);;LB㈲5k J&N,*?#H%Brؓ)c3pmخ1W}-0B$9uB3V"JY'5LxߵATUA`IU]P +*IiTIK :hy;ut[+"/$Qgٟp7~} Mj; @ 6( sߒJɫ€EDjBFLf> ^i]Zf5@ $%Z#f;`4fQ6:0)VOKY -9u~'j$X>X?TӵS 9]O7XVǟ?;$ҖZw)CPJqPWםD>$Z! $⳻-~G> <zvL"~OWĄ\YDw$x 9=P=YZqU K1"\;}VI4r(^.F=TgSC($WGN@ro\b1 S^(1rN񛆤¹x2y6^:&~m5`F5I5^תnxK '5N$Pa+vWs*JŤ~e$(Hm8UP*OC\jF'\pk,(  Eq PșFi y:htegg!d V9g6C׳$2jٳ:J^e2jJ|YMH(rbvgV-ðq;UNq %Z" )}V9|_\9;Yzހe+t8Z}4% m>OҒ2Rr), FaAee_ %MσDRDs)`$ {(=:f\=N l-6=*^}QDf,rn{O^ I;h%Omydߨ\И=BuFKt*8 *4EsWr* l]~oy9N5R7<=0ylyټj뎣Ipݹ_;Zz.6Pʫu͑\xR\LoM8lXXL ]b*'$O#Ӧ x! 51;=ozE[~Ek . _!gC]Kx*uǰ$TN@ ͈*(ܿW}1^I9L/\p }@tP؂Ac1/ y*kLAxbKIi, Y8fzx k:!D曗%KeI{+Bٝe|&Y`&|dܮղ /)%D% />KOɬU,Y/Hҕ)0F$zjD d9*H(`cz)(p+ppHS i$fO1&ȋ5Z(BBib7EX 2",, 60~As'AU&u.UR\ Ntfj-t*q%j\J5W{^90Du6`۪F̪̲+SrVFABσK{v΂(1?>g`5_ fs0% ND D ="FRet9Pu$oR>{(byؐ5!{TS&.hQߖ ~(z;ǡfSlfú#.c`1y4iȁƤDs HE93'=>`㫞$>p%#jL>7V!F)@S$a1qU ƩU@J.ގ>4Hrفt ¸iG8Dy;B}pz;NECn@0^NmC5z7MY / FɟA `=PUnxΑ"oy7l>s>& sƱoeenx?2bԥlzʩxvTƎIjԕ(4b X00<õ܁mL }Ga8‘̀ T"<$E1uéȮyzN5آ]a4%W50] CwO]ʭ W0iw';=iQ`C#_tPj̰aLCCFCN#IjhV׉T#v-&~0ƌũ6,S. Ga7Îח*ؗQ uB'~W?L3z=3\q!= gY'8Qf A]9$Xl !Ʃ}*Lc)^4/~fd^b.F/q8swa:yJ,ot:f ~R`s0h)q*`Sz`SN)$-*w0$ pיgSw9ž*`-FR K ڶ;o嘆HY;#[N7!9_((NHs#/WAOyG7e(L4"d6Zl^l3Ly+XMiNf~+ )O/j$c;¢7'gMww Vs2鋸khn ' RR6D4(ubete@VكtgNɾ0n"חrQݴќ_$v)߬JNʽEL6caRHqa3yN̝90m.;)9>{O"GH~%UNm_ [P^Kɩ=iR[QIhGR>*`&&6 6 7ɒ+R\hV2)UV[ Skс>vYb ̞L>yojp_h5";^Uå=u`w K?᭲mRdʾ$skvD~+#J@/w1򣀈JK_]{b.#_7OOIlCv"}߼~?n+v7j{Wy0JV?gn)1aԡ͛0K`< V_fԼ10E ȉB'5;K&8gϲ,8'pcӧM3:D0C~/WH{Ѹ qY ǂ;á'%x0jlz6.qۣkXXv6 281?+8&ˢgbC{?IV6ĐϵHTqrr[L=[LƔa5J[4.w.v`~;rxkܱl\-tnU\9Ńmm4'03/-U3W쥴^pUZocFo#s{?_m0Cn-',nuʻ7DKo(:Č֜s'>۸ӽ}z]叺1~U |E{AیXs~qY%DK!O~bWsK7t"HD΁Hqrx0lqU۞dLTPض7}e.\ѶiD؛V(RkC.q4&I4CDfٞьo'ʦɬן۠SX|8 ~~ByLͧ-(l> ۇof} }S<^j̛q#O+#?]xX\8)X.3V,@'JbQkĺ2P~H2NDT`D^ZL_ GdwdWGZ2BWI~(hm}1, gtu1]8`rj@JS& z5/ Eá<{0]LV]u!?8* ypl9j!\&gn/Cw s1b}+F1Rx U.N" ^2TS lF  \<(L0ihagJɭ" mއc=X $%A(YjI/*%Tb[aƇGߵM&v \PvٱfE#s4Pe ͡#! -\Bo^ ̭RV?9w--=\̖RpE ̽d%JeI$)LBn;.D<S1~*~J?bQ%HBO޶qYȄ\"zv8DpY&ƒV+UGq'yQ~~A #SA(LO989ôg}Fx]N5f8>ys?xR1%LQk4y+{/BjSpgfvs1(Z>V,3]Zhm& !^*H _J33a%W}NaQcEI."ℝ$R0T{"z_}SWOo/SFexk@{A, @X B T֦ ۸C%>rU0KJ zb c%S+F<0 Y Ps=m)edJ^Xɕ)x*~7#LM51ሖ`O_k`s+j@!|eXbBdAme4xԫ,L* T "[[&~ ?|?R*,d Hi2L0kd*hYcrA)`\:`1d#;Q;~y~a)1PP2-BdHrPAԾl5ݛJ?|P~oWTh8ܛg";Q ֲG}ll'`ZT hL87P),4v4ZkarB̌(%!Iذs<209B TQBSfD"+h˱>͆㐾8ZEv1tf6Dxd1 B7QZ0Vf8@gΐҡ! Tx-3^)f<6NutJ PŒl6Q !zz (F 'Oc?/+Rwy>B~Uə[L>0ƣzUj[ՁLޜ5g! !.Ta֯# 1Nvz>k WIɠԄR:{{]Efƛ*G2mscAc8_R ^ǗL/)6 R TT' թ ,4R0;LIr3>T;Ԁ[$2I((~TnOr`b.|k^J-<1S &ajŽ'Ks^0ʱ hE(~&[X[W_2/dghlyH%EwѱD{$B*a0T8uY.Lp-En Cw'RUngq< K&SGaX(r2A3B 5Q, %8l{W(`/TUx2*;#f qGLa+=/++mS 1Raا8bU:@*<"Y(\$ =g\A+\4%ij7Y)hzwk6alk(| Q%HVJvFŻu)Qu8)G7Őmj]9&q0_#`G5Fdk@C$3B߷ ;s>_#As NXQELgTM(LiPvZ`EV"'y[nϦ.{ Wd0 =3?qVx9(`XJ^GB@iO Un⩮3YF$F3* e,LDvGrse'N6҃3~UCɛB`}kf$q 6$t$d;_pA]]tK]>ZPXg6 C~FU( KJ8^tqD^n/hT$-+[N$L|2m="`^RHm%Jd {ɂUОS.CsC7œF LʳPȜP(΀խ+E@ rJ9@\3kW[52*y 0jfYaPAc [ Qe0&\0=NVTjS}$k)UXhDਐ9U&o%bXoւƪd2X1.0TW>2EP]?}?fH3pϣnJQZ on"<*@p`K20^(4FAAywV+H^dHn\~d.MficiH.SFgsHЎJ!' 7EzAShmo*L*t]4bG6r94FKLZ a 7 ;3x،4.A*e8_?F)./fdN|Jt.F*"Q/t/ZcPZrFY@F40b'F9+IBYw6k#hoQ$-ۄ{|՘5߳ ae6':' { T蘛e!yk8Da̫pBg ;\ OY!)rӖ+LՍbro!pLڬ zfʑ6ME9 1c]AȘ3NFkXA 3?/}3avWиl2%Fҝ_$=c}51>p߲[q˯Zu^uyޔm 31*rcČ|6+Q C_@LfQ[n \<MQYddO)Lʣ?{WF #Rm`< -[dуIQRTTUQ`[b_{01r26pf+]pbWEx%'(;W]۴sW\ 3£ڳAr){caeJ.Z^OΑ=eU\Yu`D7{Jﺎa 4oUQ*|6cطO2 Q+)P^W(+ڌ.\׉,$L=e&o, dR0TeLv P%@p˟~R;WJA|{G[F2:%3_I{͹l`ڗa2EXlڮ3W1r6\x:teeo0G>ji>6^\iUy{[rϾ>6T)lu9 ~6J$HYbH-oXQfDU8qџ}>7ӑOWOF4]L\o]F+2$MguJKzOOgg磿ӛpG_jqIONҫH=9sqRժ5h3ñYmGZiUͬ=!eo߾':dzө#sF#ё$5AIJ{u ds5ʺ`e2W2'0vPrh^e`i% Ak ݛSu\Pich [ KO_-l $QJq: lR]uP.I/1]-j%F Stqipc輚UqBvVdҪgQ+%ZG>.Ph2Yy)OJ\)HA F##GqzWǥFRR oi,H+ 'Du.tQaSߞf,*ދx$5Ǟ'7hy%t9yNITk(*ō`rTPA$Ж>~=O+.FuYvKf!Al ?U1RWgqa7Xim4c6 (&|OMet9s`+q}\4D0BnPLZ&cnm׳7K |8=?mZ#-n:~qm:qB2S'qPaVnUxrr>_*z23ٺ%_.ayDv؁\]W -sJ䥅Obq }4톀v6 Qek,4SV vlPgI52=W7 yU‡l4Cu09!k//H@Edﴈ NC ]X{/.*s=F +  Z&Ń㖾Mg8 X״a̽v{xFLuGonz9)޾GivNg:ӷ7z<}?:ohǍ"&⋊7FJa8sU  GcF#_PjgT bD$BI {NxFB}<7Z^(PZuI*ѧPD_v[vĎ}IYTh6+BW$yWvڭo9@ PF v,mU2p.(h^i(z߶nw6M?[5ŇNZrXk-b&ixFԹ n0Yeh'Jے+p-c\ F&NTf80F8`䂑W7ҕǡK`:O%8iԵ'{ʫR2(+V)C1F)96I>"A)R2yCmx.S07}3{]"uJ k=G 9:ER͡N W Jܡ,m):=#MA@frr4Ώ6<⼛=,)g1C.'$3wϛ }^yѯ亢5ͲH#‚ӣ?Hk~:=zفȊ@ŵN ouxX g\w=] <;:g[KڊFuC_Ly<@7{|Dɻz̿eU.r,owXE2'z=0d4]ZhS[X}˷B&؜N#d P=A 'Dt*2P ɐ$j᝷p`3RN%h}+ ARjUW-~X逐dUdҘ*OJ\)HAO€I-˔1M8X2|Ҝ1j%6]4csGI{>T7/9Y%Qŀj*L'-DXYV7Rsy ,=$%EoP ' n-kJ`>ч M<1m d-غRg\Wq6)挽I(i/kV1.TvӤ[:&l-JbrU&  b,]婉*")C# . 6oRa26QE cXʳhy/s횜Wof݁MVAu>2<YvҒkШK 1'ݳ4hQCr%-$/H9 0,9Ě܂ ^%ߴƆQ̑e=hn9ixg쉇{ݸr$ŷ9 dӬ:  "lK^L\dʋB)& ƾ1ZÖ5;BE8n}-TG鴞{RO @rw,[GہH]7D6 h :ֽm"- Or-sr @0z0D?jﶆ?h&9Cs`sf[ͯnqd Ď\5Qm f6C u'c+9 vsϯ#$mHT-ӄ;X$=O˗O3 Op~l -k9[~,6"9fp,/k[?)쾢KfH<.3Kγ*2/ /S"q߼U٘/e] ڦjeZ^rȱj՟վR!r+3ɩAʽS9hPO|q}D(XR.FzeQg폂KY]ſ\-\_]| /~<9eRÇ7y/g4y=u'f|vq^0ydQ#%j2)x5pNk 6§)п6Mu#)OdZUM>ؙ{lҴMʏ&MɆ @M/̲|]EEGGFƄH~ 2)ZCtS Sj A eq{хqם^;j'CS4rlVnF2w}х7J7*76t3:?f.sXp(f#ʚ94?6Xӂ"|6ޱH˜peZ.s>nǣq!9F矐?7ܦ(hj @XDQw~_%ֻ0FR""Z\B*0RhDa*RKFT5;b:G$q6*U/n)c0'NZSi PPI[,LBQ^1~ 04-E{P] D/JPE1] \L f%{KB4¦ca>(+%0gR|~Iegr5SbSQhAu{-U?n \,`LCv cB\ qY{#Rɤ+حm1wK@-2lǸ4$ x{ 5O>}i~~űm+,NWإpwG#gn1Lf 7t2)[(;G1160`[kȪ&EU,f铆+ &Q]14-klZ |rRu-RSIqcOp_jSj[jJKB1j.]+"Xh+e^o֡vnj`[ 9b:2ހq˜d3L(\O\gU0U^ c]psgoR ˳5r2GPl5+h(_R{Y8O_B+.%g ;4 Bn+ܣEH#$VS˫/zO vF A-޲-t`)?֞0p\LS[ 3aբFD{yl2Fj Ɣ8 | H>,|F#˘$24K2Tl"9ZMQmAܢG1uN73։KHpBf¸8ѻZӪXpJj^aK0b ]i shly3@'"|*V˫`Ou{fwaldS(8'6cf]]t[rƂ6̑ht]8l!% E1ᄱw[{犸 hȬʛwkrÝVh9 !GՊuE2axE[(rBܪ@ݥ^*m%cg\VV'lw+IA+zk*@i5/GoG&N* @lQ0t~˓ ť"|Q+0s? Gp<̅;={^qKgP"X[r0Qx1ivI?!!-zx[ ׏i/kx]z1BWcPW.FuC`)C[BKyJ!pO=\ @xGX^}~Qː4*V-7i'Mlw:MF<|gA[K%So: 7DͶlP O|k( (栾\ƀד2-'o~xv/UU"__-u'lG[>[A!W7i(S{&6Ͽ//_jϜ.?V앓~Og;xJ4(B.5Ѡ*uI'=zx:~u9TK`Z"MMQԏlPH1}md}u׌f5پ;U;K5 pW'-fܠ77 yeٝ%|DHѾֲ%ZL F 6GZh*frXcD )D/s`&U>ZaaTJH";P-þQ`68oor4XZ9kڞ-gRSu6T6XJ!fHqUgKX DMQ ݖVcfB&QbnljsdJ4Wt.JȨJ9Q"YCغնw=%qâbJ[ }նGfWRr,E] G,H\RBh%O==b#F?|O؉21o^7y}ba."ΞC5բ2&|rJ'F]!T7nJhF$fknMQD_z( %K (F)!%>ԥ'z]61HVr*f.myDL:i>"ѷ~"w5ּ* DZwFAFE'ryc>m>K[?odb>g )l ԑCt[OʦZ`7 "!s-- 9Ԛl_Ka@ #Ub].ފحi vT"8OՀiAڔyc/h.z_ bFbj~˝ eA3Z bЌf׷)7St:^tz=ѠS] t Ch`tz#ljg0*1H:/<ت45|.gBjWg/5iXzϼg?,SϟBNgʇAV>]0\1e9n۹5vXO7f1rGSėRVw=dH: (:Z+T!K|7?r$ȁrܓs%BkM\1r(/;=q2 г9۷sI/ē"wm;VJWz^:sCjz`"kСŻE~o!X2(Lĉ]6^@2ǂP <ږHVDOu`9?8­q1<yfup3y;J5xf-aEݼ[qL/ d9=3@p7v(veϞ ;yPDKCW?~ϯnH&jO.Ԝg; ƍ}'OXΐRgEBULCNyeV}=UYhNIg'1?Ds3FpK\B1\eUqp L6>ɐƅNChV^S߼W΢KɞCV{,էFlG?> 0Ӓ,`-~M(R'06p^N[OQ<ۢ @݄P,ˇ]K#J}\_roZ`\>tZ]x%px,rЈcd|r/5EFa-mi%YN{ggkP.JSlK/DɾapALKrAwZ>$MALA{րE}?WQ؃Cʍ<ӹG3Cwt8 oП^PQXv֟ī.}&.ʇ| 8vzc/Ɣ{ @Em~ oP qɅpJ]i%gQamIl}% چŪqmIut,N.MHc -*yO6E-ﶟEDJm."1|zJ*{o5&e&R*Lǽ,=r:;~lUT~"n48i á;sS^ ]x4^/q:xzU},qU0v#q/]S$_Lo80WtIպxTڨ\Y&2%8u% Rkqሩfsssؽ-OӢ *#0z2,hԱmථޚH H[J$%\Z9ɔeP^8B*K6喇7W|8w3 U\P@a@ag$l2P!J֞;3`) Tk<:~\8ԝ(U5b oxX^5wLT7!1a4o"ٍ \ۨ[U5HI$"{WʗlGDx;XNcæ['9ݛQ~J}TQ)LyfO k?ܐ Д2pӒ)Xӡt( 16xJP% ][>*%C=tL2<:r>ʚhž8 ;:EuѮQy`]`"'R\VJA̛jV䃍2U-_([&]]K ZEciG{芢-(qqsLg W=m߈װHӗ ~O<:X H3}y(W u@˅5v8Jź98wemKzKUW?g$r'zeK,J_"V-kC+t R9P,/ 6T=WlE2Otb`.}(v?w/TQL;ԯP0/dzx0n^y뼶iM;oA{5<#+v0 ^iͫ.؛2` rGښe3n ᧟vwGg'ٍc>~%/b0<,XLN甓3\8Pr9&Bz'vo7|w_!~RA]xK?%˼:=)wv=x9}Sp2w,@?܁u/PQW܁X(șx@fy/y`zPz0$1w㗟~S{A)>):p!;췯?^p~ml+Crņzx|U=b5QpNEd!vb7`/t'O{oX,x,Nh{f}з'[&sDf'>oHU;Բ|Njd9vivb47OOKhRpϽPQo]y3xcx}0%Xx5hǚm r6Z~x.g_rhSu%(9gwY@|Uy _F|41t͛+οsǻnM'#~8 Aj}g> jf[ՃRL!7M䐞E\!(=Nn0տ:bj<}"YC_|,>DsaU҇C'˪WpҠ- J@"29 B%ha_b f.+u!YphJ|4mqi& %SZ<dk%ݦܚLk2rvQoBSgÔy1ME޻ӣ䐍;:N!MuhOF3ESE)bCVq2}[iᦉuK&Q\jAkj`[dE1 AvakGbѨ.clը:N5 ($N!z]HY-+5"*m5b6vg w3 U@zy(s4=˨zl3Y׭^ZǬs,z%,ElnSC`u@=F?g"dORJ}>|֎&V.Q:d~sK`C$Vu @mqɃ7ZV8wќc(Ύ$46Y`x+R8!aPl[ZqU8VT`C_u:֕Gm]akJ:IB"5 z[$B\ fSڀ"|GB0Ofk=,TOE} dBQ" M}x>mkДlX|KQ 4Ƕ%ѭ}0t6EfoD.9k$#[ɃvkFj(S5l["ٱXAc IN=UnMg|,00Bm M>r8$#gWME!p|ɷ1(<ʐ$Z!Ga`HbN*!.K"iph]`K]6+~66<}]V*|aH㛌|ZOC~8{ɣTrZ 9dL w~֥΅\g\h' p*MK k9Θj&9WkVA@YX邶I8^<97w$"-՜|Hw%m89`")CdQ`JUX>g*S*yYWENh*c5SuR]ŤDr8ҥ@MȣM%@@g  1u 1+F'fB*{-$'/{%\t(:&1Pc#ǹ 1{o<1|xvpz6;Gr?H|Uu+JQ_Ff6')Y9 ^{%މt.F:0rI>g$Zc!rߥY/eطbqFG96Iά B %ͭƌMUzl:ejTzs܌Ө(hd58k6~|Lau dcSmxScݨj|m-fqJ8=wj10CE WbZ/E1O|畘bOKb C6_\96zdV1s/nCARޡw췚;LeE4\P} dcjx3tҪ"ӝSpjДd,uW huXObE-Zl/-o P6l'+Xb*cr/%дVl_WMh_RG`oxoov_`u!(~s)|[2<%YOlsĚ\-˭Y9b9_9>O$Ɗ Z+ j"Ĉ4[)1֧Ĝ-rWL+BϾbE)P;BR)ĬXv TmW*T%̐QT!JBzaB4V^kJSYmFSOށw^K29WLAH^Lz;^Rv\b ޞ^Wƽn+~<9;׼:nNv6H77!rw@gz!j^B5[͚:`2٣t8{1-*K:8=3Rz3!īNl,M\LVF-k_izuTZ@қ0Xt]RŔVm:kJbr;co(-U~D]VbђNEÈHt}'m&LoS)"ˊmԈh C)PxX#p}HBG'ayJ"#}ώI"BUW>CN%18;6 !eDoņ5ST_!qB嘽v"1r`*@D9Z5v0H|%;kS܈!i.WuAw[&o؎;NDtṿy_@Ńl2Vxuщx*B *F&[>d~@<O}K>WG6Zșr pỆ8NWސ(/8}wF0%nI̭SKQgw7t\n(ᣃ}В:`?{E!;|EH/}G:sSLay&5(w6|j1&

Rl[7j7$:Pom+-T"FWZZ,fY(N!?Yv*~o³)"2oYO>O?^;8d9_8hhgvw>w|W%ӱvdud v]|nq` 4'+&-37mpn}As= e4hdwߏ:bk T_0Na^[^q7fu񞺝ou^>Ynk ׈=}^<W|pt .Z^vqnEI 3)<::y,?QQ⣯Eı(~N;8Vcb ?5MB'w_wd'O>֭vebz9_!2;SR^a5x1 }XK6u4I1YJdXdjJ*vK/"ȈM̲̮JJix㙉yJ^:5:gNGTȭ!]1]J*=*G͒&H[Yo<b RPS]V,q3$E޲QI3$@`},;<%C9eKy삙Msb!UNbQ)ه+Iug"f0*2ThƀZEQH 9-6w^B&Ijւ\qڵJ8%Ri]$E2K>ʤ*E~wJ j _Td'Zҫ+D&a&ĤIoDi 1']RJQ8 N5#l8Jb(1 *Jٚ@Vi 5-8U8h$E!jw#ke ]'RXK-ҭ."7-9=PUNJ,Ypz|k5;1Һ8lGO +Mv=N.fs1$z:6>+D M. <(8c=XI(TF{cA&VǦEt^d<-V3le@t)t 5+@꾼0I_t$-%0γPeZ(61_7BnkZd/1*-xy'MHZ.,^y68bU=>[ K +4L- 5 t ?ǻ4{IPvEn"]Cڪv|EWp'JE<$⡕a]'U#*zxp`"]ģZiӎxhM/)qpdJ/vD!1pK EuPNK)ktMPFӘ;HY)T ޫtcŘRդ@KvdG޴*%SEՋU^bN؋^/*h[ E ~sy1]ijw)T x$^SkBsubXTb𭯨,a>Zkƙ肁*NE6qPIhԷg^5]D Sس!wF,ͺ!+Q ,8(H]t5x"ҏM'S~{L7Ipu;+:Qpc>Q\>Ѽ/J!4k+XCK8y "[Q@CQ;P܎2Iw2{'S<g> 3~1u'/dv'ib ش_dQz$o5AiZI垩Й B{F3PDk'Ic {:O߭$+].ۇ=؊BiV-KM6A>=>mR;?&ZBmՇ+;}f=fPXSyt+WK =û>Q}(尹o3r߾er#VѽЄ^wYiC" LϦ|{,\0̳ob7z'W?Irp=}X_{_ݣWH$V Bv*QI-+Nј ;IH=#KE3C@oGk3xJBCcS\s_ou | U/ (MbWUU\hÚ4] LRM޴ˇ;7U\ݻI}&@Ӎͧv&!U'6￵r%,OSE*\- ǽNH~0 V3Z QA%iuӂ8tK`oW3.CH2d!bBLkP2 e!DGUh%|`\mj4QnfB(PݙawfX-ʅʺӌ&2n,yr0|;jhmjX;^ӱ,4.h5h(UVbh]^ ;yk:辘3 u1g89GsT-9zQSHuH:OhKu<^۳N#űULzm|JErI\FQΫA;ƁA%:dp,t^XPNءtF6emq#]Qk~/rRKIj$y1u2/AEP%/~US>:,t_IT'}sbɆ @R.X65 A, zv8!M ] V7m[fO^*A59uy*nB ftu_׋>ɥ\ >jIIx]Ѓ8r W uKak]O58I:?_ektgp$(DŽTqUN.ZOD hSDzͽ{dZaHr \!gxr>}@pߋY@#(_i^2i) <%65+qP4uyr0h"xΩy6(B<O'Y]NR ⦅CD(IRh쭳_PxDd:k hMCqg-T)(׆YrEЀ=}#G'_kV^ q\!y6)j$ ɉ2Zpk U';8qf$7[قqÖuaEtɄ,`cf C e+lD)ȂaBYRу10r4f#1񄈈`퇍$G;gl8}uP#s&J{'Z:١[ '"~xKqŃDTM]Qn 5nFKgڰi9%x& V UDQ3!p{be.P$$W<8(Q4oL 6jFiVls+5X.Dc= YH~E1ɭ2mhavxBgE#$#`zĒɧU iE,5EQFmFby+ȣugdY wrJZK)ԗ8DMr>An@yv KJ  80sڂV (#9SݝݏSD\h+ڝu'bMOHs˲'W, -ytDѐ[ rnuw8; ڬý(t0`X([jJߛB4C*aVv;('n{U^'t0Qsr5ܡL4{y<~Tb.s ;7|`2{[eF7(};;*TΛC}/͈ޔDJx'rM8 zYy9)V|?gh|eux| .y$&%n'`%䗑|mbuib8WSMc_YTÃ*$d w~VO_]ɤw3a1li"$S %ILza06jd&-ki p}*Kׂ]mTl0! -=uc8l BGqD)mck:~z(3<%>K䥈S&'!CNnr)O1Cx\Uv7tg6[߫h//4 ` 81ˇ (I0$JɑnPp6C @MxB5R+"& D~*O i?dOUqXL )~ 5ol2x#Ƃ|*KF7RwrVu^׻H|d}%/?;VfZJu0LJX/_a(Ǵ .RXHčfQ"%*t Z~Z7 ~v6ه8<<1 )a(If%[~^.Jm}!P'$ql"hUk+PD)vL/SORor\/*h`@b5sbIKVml]~ޝLQy#[=ޖR2#8br.Vv-sBZkVx·R\SUSWUy*>>Y0MYw%yOD#l "뗙7T7 uڝ h88}E{_a39¸F X 0&.y ~:\bJg#;@0淞`IZVvσк RFDrntYlP~q0<+Ě z/z' 8-6\2qk%foX;:6 ߵ:~0FO;~e )s6]+#&(pɏױu*&K4f_e6[٧o&1/ NLl ǂ{#1Xl*Ȕu$*E?vatS@z+,Z@J7B2R ዪ 3a!PVSTqlA-aCP:gK -$Q#Eh0i0 j).*1rqbGsn| L[eOn`0\e?߼y?7pܛv#p:~|F!Vȿ|pэ<` 9Ƈ+Smg"-qt hw/]U}be()#Fb␀MGcI-E'DCyǃܳw _hޤrfLU!bNCA^qִAxñL8 5ꝃœK[#%=xwxWY(rK?ˤsgF9{:~p7'[dz:\^O\zXٵkvڗY˝}x|ru4!ѩ1^=qՍokXQ5q#gxCo>KۛgAtRo{A;pQ!;#`?]b7DG#@x@v_d6#"EUE1`DͭT3X;؋P˘q0c a]xc7hv@ދ$h n!5;:;I/O" p<ꦯ}&0pBs0##Q(8LֈFx ՌץF@dC1H,kŸZ>m%x!J# =sܚHs^˜Ͳk wR]x9tu(XY*S7F> C% -$7O[AOy«YyقO J;O;};0GL/pҀ0)Ö.A(В2L$֭./Kژ?iT\:Aͩ|6SK߁ZMѠ&_}&0s݉ۥ&!3E)bTXDb-~j^M`LWyI-/0F7}_~\mx!"Tfi Ҽ~]''ՙ:ʓzyvpTKId~' g|m؏~O%5`סVBE>@4wXDs \@Ät& #% (MyD,-Ü'ToUJnKN:Qj8ʦ;ʓZ/>ش}kp3P ǖB#LD]+5ZD]a)#.J ?Khclq[nS#1DaLF֮VV|ffR/tiUzS*x*=|Ϭ/bxK^^t0}/E`p'_EfpP5n-WP&=dWc\}tőW&ɯ}ek_X8@VKH$`Љ$6\Y"8N$yhyhpiQuJa8>+D3D%q5.^^Q> Ml,̍(OPET+KA̞wE+ _{*>E,l66G-;'EI[/li<cG&yX")Lo>g.fw!qzjPFKc\bC9lɣ-*-I޼'+$Kx1HmmmĶ|r`1*ٱe6+O\8PTRJ ɉȫ)˻rjT8f[",x]|WH:*<0ޫr4p0f^2zNu/aקTՙ:nzJ?יv ǻ8n>GM>mQm>-Ne!\k璡40`J9A1%Չ@}H'<%ŕxtD(&?֍$$IV( PIbyDJy)IFPA#A_$ g`듡CeP^`kpܩC_ôNE*ư:E; ^rKEyBIQh^y}|*5NAb*27hx.8$Og<0գ?|^O|Qŵ{o[&t7Q:l&_X+euK- i҂JCks2 -Kg\Xj"Klyd I61%kze@%^Ys&+$PNH2ws8 =J e'/:%n`ARᵇ|'eD 3ьc¸@_@9 5;<*JkbQϋxDM? fǑ~Έbit<` NJfRHHAxLl\ɇG#TCCP ѐ25L烢hBju(q}Im&lbɑ:W68Q 8zE5OFNE mHC DU ֵf-UZ.6bGLR@F ,~'0@v1ϵ* F8N1nj ~y:qfNzI#623Fg. G; ; f&Ԁvτ e(ѫD!P"a.3A3耦8a E5Wm"yWb!9"WfDu8ԟul<2x2v!!脵&zzʢ鸃`DIB$-U)[Kٚ\R]ʶᝇՉcfLj0H bO4$0^X;DHJOZ. 7J;W{>Dw*KkP$sùu}#謲={̭'9C~[ ۈQ[!rLw!qRh+9#f5ÑdvF^G~๿YruZ `{n`hPm1΍'w1J!qΊ,®OV$eEBUVd'= sVdeΊ4/%N74Gs8E ϫ晭ɬ y}խwW-A5I7CZ5*g"l*yYm}ˋAm;O!1hֲ<ٯmp;,Ũ_ppz :0ȀSbhr P1jR2 E%w(/ug:zQe$Xzd]sۄitLN+y C?_E`?fww۱=5 Z\U9%Y7;d RsMBfEFim$z Rh>PKם ЙK[eը^%:.NS/B\, *fOm4eA;S[/sXiOفO@@GySGzd@&lI/If`y>C5. G_ !j3̹m!i1 E;wH 7Z}oϔTBقqNN0 fa=Oe|"zN$mDFsft>4[o$~J޾`9m(D V:h20LovUdK0;v@bvDmT(Vyn",(S1?k n]w!h|dv}eݑ:2Zq/8LS.,p_e} U6Wmݦpqv,dD}QNIbC^YҎ%1^½* 7D 5U ƒdírASNC)SS>{%'aʒo̲Ƿ;̉jUxqS1[q+)~|YX|~{]If $'Q"5i-%~?JP$Ӝ C)9šwA\ ) 6tL0s*Gn#ˏW ̩wB$/4Gyݲ%ܡYtryUOnnf.'pN)ېr)x~n Xh k突4^OјNrURTHJs~) wM$4k,DÄfOhɺŋavZ+i4^';ae |X!sy<d*ʆ&?~$x4Msh ;Xp/.{q$,w麒W_ |q"j@-|pT]#UwpA |p;I៽)"Z$ *#Mp7_FFj4;4kхWz3Mđӟ3]2h`GTliShńj}ȹ 1'G^Zhk G `%)% L6:(V.CAcQt ןaDRKuJ.=&e7VH{ P@. %b73f@ t 3I*B+ܧtXr:ҡ4t2)ҩ:* qf= >ېnRWmCv )Ό{y<ԗbmGs3\)r@nqٺ []'^kB V:-z.RϜ+_j ]B^l<5ZT}J<V%JC$,::=fdWbFHUТ< ;BSw#5m/®nvSwuR]~.Ws!>^GMh]5IآϪ$%Fe2!RmݠUD8XB"B"4A[D3FjMbG1ZQiB YBB'V^8.C$F57,|=KkTD0~j.Tb#ʔQtg8j|c^%̔v)B-dX=|"Ԏ2cԈSt)j_]POăvpު/c9>Ca_?Vwsvfn}-ؐ@$aX꣎ ΏV=ݩPi$w:V3>+Ďe 9 >s!V5:4笲v;>7Mnr2.`v^<$ШS MkuY=syY}V5>[\^R2fT'AG_. woFaboF~B~A2ӻyVuFpFރBy`bonD0h.hc깥Z{Iu: E!"i!ϮQv}\#Q*רk O)yagA9IL<,N.['W9DsH!Y5:[ZةB2cp9Y1qߠT7ײ4DoOc|y6_1%z5jG9Wb#3h>YlfG3#lf#砞iJ Qd#yjnT7,K66la@&lI/IEP%•b0[ ^Kr_-)'E})G.-~'5ў0 6x[Wlrr\`O:&UJڢf9t#N؀B*'zZkZL0J .D0BT<hvN0dghei_ G2[ ކ'YfzgWIX[^qț>d>䩿Mݷly5 ݻn*ۂl -jۂ{}5>N3tQp91nW?`jy*UV˫Z[b\D (Q5LWn0By(+@EuRo+3(?K{HǪԱ¤+9lH209BqC[Tfy8rFJF?˚壚}^>D0WW-hBTc)#tRDPvp3Eᜈ. WKQ8hKړzL^'Հ6vnuhW`TcSNd-vbkmq,yf9b3%HRu<"U5q@8U-t[o62ոDLW}x\yY<@יٳ”`U}SwiŲKkIfyOlV/f2삤:ܶoVeH%+O(MqIxH~۠%̔P~);F%ua#?LV~B3?N?mUk` 'wOGBoyu~Z\ ^Ng@z&~y#uFϧ9YoO7pG;IjܞGgdz0N3Y|wK~ȗ׏~g_:\/Foٿ޷߽k~wl{lArih0?'!vӝ}{z;3hܹ1[ .œ?X2"} :~؁G_T1gpъdN7ťHN>}0^ ܙ>27wi¼Vt;]տx:Tt۱?Bi+Qis3O:8uh(ѫwKƌBkxf*gYtqwSziHMr;/LI>BiuL:pPF({ 2]L?ӿ@4 W m[yly|Ot<~fGl+fgȌ^O%dfni >ۿ.Aw?SJV @#4"RmwbEk50싯a؃%!^ݖFJ1E6q yGV=N6݇y"1b/Q5/b0Oˋm>{+20od9,ehO:`H!̀o>We<6cL:1l;3ѵL5QʽS XZVlJ<#Z<Z<~=exܷjҶGQ)AUlHE`zG!1Z$O5r Y (g+0X@?dCS!zЀVZBL  rP`O[o ^wYs_n{rwF5Fr&%FM&dQOl!wO?^4)aÑ8' :~Ƿ'G뫿T. &julۊQ'N y4I{ӱ7qi0Ş$ 5wl]IlV'-Zv)L튟)*X׹cFFC9{ER,Czح'[6\FzIV7dΘ (8WO輛{_mqgPCO+D þ`z~̷(Frc||~g|o]I'n8' jhczA؍+0 !B/~OMwY\Z9 e,xƒ&'<RY=S Is/Iʹn|HLvDs~y hr~NL׉]xgo #ogw[^07˝,A9'{ :[HQ)RTDAgUU>u՗M\4qܸC>n i0ԌG oݼG:X!Rd &V]<z) ÙKŘaZd%iK6㡼b.蒍.K>cO_(GD3Q0$bbizi$IYK-.^J`C p |#LG,~E-6tЗY\5߿MƜh̉ ̉/~Ә9јj;yI\8>aIӁi,oFXƛ뚚,M^6Wd)^Y[cp4kN. .?vV)BÓj#Nr/gbcz9^\F0)EyfN(k[ˊXWu黼uhN=|0d𵿺ABr~ tC(=S:;]S*A~y+FpzxүsȰJk'^Ý8gn@81-:hW? Z_.m;vy+avWneVq,N71c9 fcċdg:NIPnlru]b9uinFd,'ꮰvTGڄbUuН&;CT7]s;(b5uE۝NWU+I&: prg5&7 O$Ubkv?+$|sޱòXusjٛR62iO' {lJN2m%5KD'VADغ FH%<`Thw[rԨwy7=`ZXn[w̬œM\oXϲrܵAr{>_o?4)HSvm+w'IQIjiS's.Nn:K{c0V;`Uء୾{ ZiA:%È *צ5v`ةĢq+ׅK5r3p7աFURT RڕDY5 0+ jahU %͜4Fi . ᤺3*5od~O{0BfgJc?^K'W^ ַSwCFd$\$Nh=&""/a/u@ͥD Qg\DŽH$9$g86Ś!ꇑB$Rb?Pmgnŝ~tB-[\zb>:Mw&W&Ҥd'*W*0RYwXps%R"%2y1dp=L^Z{iYY "2D ʰ H)# cB Rsa$ẍA9ǜyJlٗo%P&lKm!_ ޞs;+㦝&5n7TĬV$"_0#ibW8 "R+<(?Bq1J0UĄc::`n;Ad^2$`$g43"1cz.%x/}ȫo{hgW>޽^qT eʀg*U2S_V`"u(2&C,}|Ib0PT@ 8+GEϙeRT:\Oc=dE,T!ar-t@$e:"Xˀ*$ה>P$bCUOyE{C&ac, s#i1)4 Fp,8"+1j3jJX`D_Pϥp7wiuN(D!pG1F`u0IeBC"nR((h&RfJ=Z{NPj%n:ЂQ ;tsqdA(:aTPG,(g=g1脌2Kk}jKS3~wyk8J7W@...kbȶ] dr;ޢ/@xq@U4QM̋phC eP TKXj]:l`vx6z\9!! g%h z: Q$ eKX(KUFD=x("D]Dj= `" FLkz  (U Zc.>,$2#=GNhTӻ[u?J0|v>{+ !Srz2~/s8y1h}mU#Qo>iZ[?:Ɠm<Jk!N'37)]mI;cK ![{o0"\hAh&sg"ɘ/!UV[-7\HF2k54a5VT!1Ոcc]>\I&;9=;]Odyɼ;}sOLQMCfCm hXhlL˟8yW Lxizo4\8\Vb:$KβaP{6r$EᗍR@Ȅc=yڝ؎흧  [H=_? ERBHɭ>lE@^ȣ {LαzHa̡c HIf,lKS`@Jmd4Jo $e6e$8 r[gסj߆ kx.sw87`DX=HЬz?$,3_dd`Spo,ـ74Z swNPW"vGRYnGq(}.5Iz3rw\wKA͍C*r\IMݐ /`$DՎ8ǒՏ\ʰݶU̝?T!Ŕ2ƋJn%qt@dPh,NL [ԊXAp]ikqeQneu + V[c.=)?YKm]-{q,_~Nw''*۲ڴSFoxx+ ?zW)HV]_Oޗ,EQ-K㋜_T5_䛫O,fK֬73{lyG+y[?WyZD~6ZR,;׃w4N`{o\ ƍWBuPkF<.蘭 ffy8и1ŶBZ쵊hk^Z)[[1c}et5wGYijn]E~?Dl׏?.uҵzwB;+|콞 d}_ָ|ZO˹5H|('##{=mO ;qOWP:~3WIJ)7{UXSkg{}wWZdFْӖeLL|i|uxIh' cmHhNW^JYt!k>|U7Ǧ| ǀYaT \+ʂ:kPkuKZBnQ[TXuB6 lh!/g=^EX|?{a 6z߿|qT&ōߴN9%yy۳$x Těl[3 VQq PUVTQr Ԃ\zlإk5$gweRYBC5H;ue>NX@%&8~ m4x ,j@h^$%ר2I) 7ӔӬ :DyLfg (+9R+$2.@yAcZd[eۭZ{[L:QGT JLGFjz-lfW=XMnB,rLJcMT1XFjnM-FYh5?o~=ƚP53sN lX3F,H̕a-+mbXs|ƑonInaSD-5+Qغ>(ٖnQN)ddrdPV4SYcs%ˠc^%T摲A=ί* a,sax}[/IV󻠣Vܲ3Y^w n[Lv[ۃHOVabGD,NgiCX"Ngdixs{qhzrc<͌C6Dèv%nf#\nVd"s@6[ N)T [:c͆2ZM+=u|5U3F$5|s(}7X?:j!>rψ{Wk0u8A{7CA tLCPd!xϨ1l)Pifb"m4&EM6Fz>,DS Z桩OH_q@ݛ`?3=Sn҇ۻfn?cow&ݻG?wl߮.NVjOް |Ȃ<=^γVf.G` DJH9T `X[{X0dQ"ƮUԲfi#Tg`Ì-]-+xPL7Ggc\[ݲI u݃.-;x}Ȫv۩l2 |НM})A@ҍj%.Y^hK +"!4^D A R*-j?_ЈrـԴwk_swɽKC଒%?-|W*A+Xñ3;<;Q SWG+ɾ`3s2{`+,UM-fhth$af lu!;4ϰ`cUwÚF(Kt,xd\;5Pr"/.ғo{iLֽ\@TOd ,}?^}Xe>27Vh5hfnq}Sɢ ;|Q"SV ;<wp*v7pjr𔣃W&Dpp3hJıg_`鉜4|K22 Jf!QsR BV"֣Ӎ6jдqncc 5V~܀ِ]6"ᾏJ)SXc`:te6y%V^ipwmJvBS:*ŲƢ$?a0-$2䃣wMXQv4"oE'|i|uO][ޜ~7/իO UwK7 ss?SUlzm.?u4c,Gr3b3c gQ99j4qfGHf$I4W: a gF Ulk~h4dN-(Vr(<:Y EDu:ѡntT(aؼ[l)u=7ڍFRw5 ͧ=s;d|3csZӤlUp >Ɖͮ;dгlMGfd`F7XPS;A@<̈WG#%lY y~֪Voi DWC`p;$yZnDi?9~{9s0P4/|kx jYXNCqD74!=+4Mׄ3S~٣[ f Og\0Љ^8y>46٦iTTwe+ q(|I Y &Kf!BN(ۛC W.mwf-Ai~e$ۍ2CP([HˊBk/G$j;ӭ8}!_t+9ЪGF7(ݲƙPKzf-1lWvu@0걖z&Zw+=Po8$uCvxnC$[$U PC4kFJc!!l %ٝNw;!R?/7)>qz3HvlJY,VA@lvdWl7wtb,v.]QUjJ c-|||RIxmm~ڗ 3WW'*ate+m('#H7D%Fcy HbNI,D']"ۑ˥pWHʂߘOjmFCp @,@GvO[=MVogה ϷzIeFAY.^AU%.%SQҖg˿sCJ]͞Jmr5ٿe76~hc~7?FKaڼUGڢ[,G)T>FN4honKws$v>٦nT" 4CSkNq:zC@aF|e`3u|݆3 $nތj3>E}c8Pf~E w0t͛Q[bޜs3@;wिI055j x-!*V-r,dBT욖reGT;ft()z̊yޔ-pcFTÎov5j".V}ӳB|y-4hNV`~Co <[7,rZQ_|Gx5b$%^/C޽a+v^}Xx6.Z=h+"1f)#ר;͈Khm82E=ScXrXFHp3 ~c8e/̓V'9Ϋ}hH#6v&w;F42X!}A>FPrazs%CXSl$:a?-#%Ug[1F{KPz銏XmG'8:!JsWvvji%bB2#~uO$:y %S*# chҨb$*-8K38TZ)#QeGUcTHP9QD 90^ >i_T^U^4^qTi1hpTiqcn uW7ED^[ZA8E+Bp4p ;Klp\d^\A@5[n59#6NrA7s1hwSuLvl_Q[n͇˞*y!g^{aY![Эs^=>}򛶒 `S23Z}en~SƳt4Z_nZl :eLi}u6pz?~|<7 b$P S 8!$Y8DB1UmX'}8 RtY`$[MlESr-( cFp lYĠ-$ M&L4yo M9? >60$oޞOka}Cw5_mWj8I0#x>- i~]Ci6A4o)^b]maf%$c9v˙w5WMV)/ -,`BJtnE0@` ^de9OGx*|\3>:=~!7+ r~:+"h8;766ʡDYˈby&0+2VWMN14h;i٬98S10٧VYW}uMLhPOt_ڠuUgu:1b&}A=(`d`P{܈Pư˺Pzn{MT^?ПOGB7^+~d~+dMW)Enn/yeJVpPL {m ^L g{/}<ңUɡV$=@0 %,EKH.x[%x2$b\6^ 㯤Bܾ`AÀvG);>K&:\sV$ .xRL%f 5(.q%.y_Qv{&!/ I ;~umX"ڰ*/Uh Л 8Ry4||.◻߾<:4qzټ3#rq4뙟=m]V6|6r;a}bFªbTc,{G,ZݠNގXw=@m3T_j8آ ZRw{l[i[7oG'G8#*1qqMocBq2I͌LC'W|P:?L ޓ;-oB.w+{n R~PY3KR $ (i]IԐҒ.jUqZ"'G6s mO m z7eomE5'#81TSn6jAOR4SjY-JA渮!5dHrgì$Z=Ģ.9€ D!b!ڢT1Jp" 0ZDҪ"DdR.YEjUPqY@ z}=Ec!OkCC51LKy7ҌL\zV*ZLOd(w ͓-t_1bc^7h \%G>ДB_TD{l(On:=-V?^5VVeGP<804r]vM∝q%TB*j Ǫ*! Vy K`9]΁վ&P=+^@ (F5aڜ*-!,ǥR5,T8B)!*Pr\3HO >kxbA\zSէ6*VX\YL `}Vʽ}Ӂ BD*kShN IAz}԰@CQ5>M5ӇzՈ2y6X!UVżv[(#g?vO&CA=2xv0hk2FOʧ3߯_Rm$K-zEZwjb?{bJ_|\W+0kLnt˿3V`}'h_Ƌ"EFR~oʵߩOn4a "&X xTVMHwY! j}ipsi3y;{o$_MgG~O6 r4#,׺?R/A/j>fo2XAR3snZ16!-Y XR P^(5d _T˜iEl9nσ O]j֭S';GW;zM5Ne;J#]m!Pۧ$\ՙ薊{p% ڲt%A{poTMs/ V Je&kYy@'E9(R֚L[(#YOY 楒'>sB9Tu!E5d(Xo^U,rƩT3BJVZo*+rГXoU;a|/ uyzsrA:ϑ-Вmq}an;rfOhD wkpmy-=@y#:ȼ W%w6r=7ż%!mea·^*d'3S)gM y|m064/cP'nA P!$фB\>tnu,>gJ'1fﺁm E`}qQ Zxn 4vyc&cg |͈Um)l5d/o8>#llh5OC:=NwItҾ? P0"ݟoYߟjG'8O!*1hsIw%o/a0 C.XfR.MŠ?v{])2(!#:m"9܀2ep :.(2W*.ZZx\xG"p9pLgY+19ˎRbTF+w0QfO|B%~T_൵SN|cxGe5S1{֩ n׫@Yb IQÁppG R8ۣ !@3H:צij-KY6\ q*Zd'd:tL Zo/uld -ў*kD뀁]uf--eSP\s4#GI܍)w7lT! 24 '?96G+(dVb)5XEΩPY)ITr` !(90y.BJJ#еya m<̖mw2}0b{Gdw_ě;"ň.pFkGHW" 'aD D.Jͼ$y 5Ay E5{WtEĉ5޹Xf[%VWADY --T3U@g|E~R_ ,߹ ׈LxJS}DJܳTbu xZWpBH 80@ F, Sg|z.9/ֽ-/?߸>vz|6*23ZdG<xH>Wm-2$>>?VzXM8 ! [ 6!I=Of g/틿yVLfUd_~^\orN|e;H"糙 E_ҋW.:_7adU@ \٤  ;E-E}eez NڡTtFY'*Iϩ?Φv̑_LԸxom][r+}H:ݹ_XQ*.YvTVY_j0]\ɊJ==(di{gzAhHcB\7O>0?\žYq% F< 7(G28*'f|1>|߷,Gʹ]~ϭE ]XtNTW`BS ,Ҹsٚ;[SV"tu:c;esWBT[+ lEw,-6ryVPC$q ]lѓp푛^ʪn/5SzH%Z_ !_z\ە\fy&r:Y_kR;R:n4Uۤ:mm+֦%^ Vc[(w1ŀqbjRT'zt1rO-#wY<WS@^ӾvYcl*ү˂b[6Q2q"e*OG.K(QLGnsxܼK]&@Q\V&H7 eEbo8ȭ5n""G֘g w06xM ٣F0͹ֺ˼KJu%A`֝xZNQ4 GiZ$IB N&\pE &(ptp '>H }uJ'}<1|f}r7%TJu.,Sыw,}2ȹSn\X"˖?A\R͹4kY{-]ݲY x aF|쮳cQܴS]uV~#%T@+μp}7ڥ"ewh/RDJݭ;88F*cn:sU]M5!-fAܠV7.>C%eNSNDvx36ǝd|NT0WvOniw!!M&\1/]ijijٮY8zf"#ǃV:"%qe 1)) V틣'l)##q8f v9jg) 1T |~8KO&MZ8$>a Ru#ŽaVx F3FPK`_4h ':ZŌ,8$ke#X8V YǜX$m@)ߋBjEYmDc  Zna MJ+_Kl^:, 7@%H)5: h`-WDx6FȰ8L2\,@ĀfNDDLIIl&yᆌ|*e QRJޡKP.?~M|.kڏbs?mc|sfgtM${y~ W ˏ_W-?Z}>MX?Q<_\ .Po+'kz3Ow_\?zwf:~1h_!ml Fm^}o -|03{sg.,v"@q,8xN˵  Y?2lb! AHi.]1c{L-5FmU  vc"NXp>ZiLt4_Y+ٸL`H&+W3Ets #VăJfbkyi cK*Dؐ؞an0AMS #|Q+;E;ݥ.w:9YmEZgQ-Lc^?4$w^t~;O;?.6߇<㪷̖ǼBwd#OKuOg''޹ket6ȧ4}Z+i)iS?Τ'YnHBr%S+iTn)֭.ʈNwTns#9Ilۺ&ukBBr#Sǹn(fE#Z<(#:uQź1Lsm[ڄnMHW.k2KobSnw󹴜2OؑCJ$0{0iƸ BJFu:UfaYxIUGnAOHr)˄Z~tHNiG5T76+np ,!Yg!g)Ela; kQHkxLycޣq$`$K$ęIT*D5Q!4O c^ۊƿ-|" ώGܾN{H8ppqma+gA\KR5%ŢIu; ( ;9ϸb"Ɓ&* NfSM D{B ,;24v%IKIφwU[u2C,X';EFXI$A࠼YcHn<g*c. Rpx /$>X'VGf-ac~D8A>ƪ@#~N pPn+85@IOZHʎ L= =*S_b`Ao=Z lQru-:l%[ʥٔ>)_NEvAMS~ U-r}%bumJpk'EX$y\OXK1f W}\Rߣ`EgXL؆ 3ugBIeٙ% :_W|Jf:̴0VGfLK1vc"F4,8I["_I2|ͷĢgٺ)'"Ljm3Ӥ7l6nocuEq3͞`n!?맷;b&Agfn07\`ǥiCog?? Իn7LB5>g|k>|tee}vX*5=skۖbDQ&xo&YRodJ-i1!Ea\ԧ:-0.<ไϱVM%j K ȖDHK,lgx.߾״:"\Du>n1E 8.6 N¸\l}oZ/VFd4KdiZ$|ZRDg~d?L3oc,91a\ߴ9f\֞_^??}* VkEXX]cDgt}3Yw0{?'~hfZ2k<-b o^'S`l=eKF½ۛǴ駵7~ 6SR~:5w*A979Ag9! ];OA׈~{C@D O4X&Xfj%1G ' ` k\7;d{۔ T?=A"@-zџMkMoKL1^7q+d&tc^ cOɃ\g~iȥ~QZ#vȐ^;5Y.b.bNӚ'O QBNiƫ-QWsl yM$f]J`YpH;oH^!D=9c.jB o5HNkdfSΎ<`ZiELks )Ls8^Y_3Tq~qfe&!Єb$f.; vf]+?E՟wEn?vKouf<)DNߢ?ҩc+4Lo]\ ##ׯ(/9w@W}$!5htbcNesp}ȿ(*! Y.$?,;;= \GM,8h[yXpq~#eӺA>\QKv&szͪ*Gȋ+Qs.RU++l3&v1.+#x-^Z W8"SLێ}'@;I[{(s'Ԣ5Zb!C7/3 ˫N{u.ov፻+ +!$DbMcG9[*Q;&tཀk, ؍ &B X%(N磓^Zbqva&d>V8 5h-Fh>pYЇ-cW;wG.N$s&=ɤr16#!_֐)-:sݺKw 2ؙEc< *Xe{1TJId2@EѪȸJ7A[n4aFτ3I`D}[hOI|Ti)߂t 6(N#+_Ex7;Hƭt[0ҭY4zBNrj>n(Bi:ta~ 魊-݂n}p/΢yJ \Ms²I-bXm_խ-jcIvuyDZ4jP(O1#R}ufB"? z^+3\Xĵ/ml(&ӺIiCdI;/킉Zq^}hy+JQ;|(qD)ԜSS NfbTHzxZlGUeid__^uM|VyZ?9vOmc-{6 Z+7#,xGz^v&epN-5Y5%RSsKMDMIM<wjn℘H9}śois @s.jclA̿SqX} ~-k [Rm*>u"u"t TSrwL)VL3$)Ny3TPB?%lWфQpWü!d҈9`RjKfjFޤ~&!e(_ߚ Lh-o!i$29RJ\k Gd. yYid#`6vҜEɌY M9? MGov:*ͮ E BhEhw yAI!ta>daф4%)AcX7:ۡoYu2OY#({W ޿YeByH[BJ3 )8+2d@,I$^\ ~8TqC̔(3yAEY,vA08-ɧM>mwq̌VSD"pB4creaTфH&+ZZP:$ྦBB`^OUV&N: RJ!iLZfL**z#>#7_2 㝍6vF 7;.\Cmݭ!vwpj= O$[C37fF z()FDح jwߝ [Q@c*~ ǍOP?qk(ťrl\ ELac*I4QA#1&P 6-NMJ4ϩA SpG*E$´cw0GPKOha[rDgTT?yކZ3&t}H«%oQ*$kk%U(2kjHd1`9+e4IPeO .a[ Ip3gBO+z!vW0 Ap=fqFҌ&d9W{P9\4L޶ lp?^Q:-FGc9lxh4r]@YUzzzG'a?IHYJZ$4SDegiB'ԥi-n?c/$ŚG@z7M'8tmfi4E=鄺3\HO @Do\HB")J3"ReM ! F+gVaϝ̭j79] %8[Luӆj"Hgl3X\85ca36M$R8&7x>dV&٬1©S&{jdwMOӎp^m?ǻ3ߺfVG8OX<}qfKJvK-P eTڌ[;~ |],ǟ$_mޛA/~,ӤOf ;Ɗ{An]AoXkVn`xBӇzaw₫.p[N;kIձ=Nƌj$tg_~3^Pfk&da&K5sO?~,k"$p/Ӧ!gRXZY 72t(oЬ؁]毅"K$ݡf2@#6 ̫| l:[.g|:ZvhN|솰Z pnW8a1ٌ댁Ġ#>{%kMZ;O&@bJ# j aiR;Vš6˻LiTsOz6!lU׉?J.k|G&u=Z!8 kH#ؼ@K0u٬`u%jJXn9u.^K%8O {w1ϻc1XcI5)M=_}Ƈ„!N8f.?BEud-|#*ƅݥY^٣N^*젯",A{p I"6?.,nF%ϓ =hP*E ^nPo&j0vjLms~Q| 8(4@M^~ʠfOA1ƨAvd %ڳE %Hفڒ4iOyjoe!Ú>ϵ0TJ=|pC=V 5U2*~ysAq#hp"LH.:2У@cLQsnU 0Tp `H:/{"Ґ(=0O' _UIZU tAw5A@V6'Z]{v|@hA4_|&S XH.;|ڔlGh%deuҏ@`=;5(di{&kIa7yuGA~;)0wEmkxbNeSLn fRCI3$i0; ؍VX r/wsyN'T;T p,j2Ao3uɯǐ3M4^y.V7xpB$^ i5p( OR`W Ixcg)ʐo-S[)[ +7U_iCd_o]t,"6isY@դ;˽w$v:pWɁJedT"ȸ&1S.&RIݿznxGҕM^&U5{k%%|ƈ+|?h~@Y=|2=5p ԘAhk;h2ko~YʴFrve !* @k)stqzf+'+'k9H )":B _x&C1Zϯ9I;9ё*w}fYZhbůw_\ ,u=ާiN C#VČ+cpIľ6yrr8ԙP٨29h(Ei7#/O6FvevefH`q-r aTGJp0yYIC̔1F:_qZ;*qr#~ @J30 K(][s7+,=mjWaҦR]v-t$RKR )rx1u925CntDiU|Tm$>=+S y듸"4q|:G=Wjo'nIïrUX;5BTie1(.]/gqjY8At/vO!tVE?%TB߼ Sohd)}!|t$uYSIE2T񵓇߇STSK {`'1z,,BX>NFN`@7l@~37,CMP OstdJ"o^sI؂6`~\_==pV|b55}HCD DA>Yog[<\ j;3k4>gٻq8 Ѷbl jl[{΍ G4A ta|娅p|m{ 6=f;9H9ctcgcϟwsAjYa=z 6 }[?tmkcyjO BPFg = pSmOQWm=)DkI'BIKbOdsDź5I&Xf l!prqf;9^KL-ێԯptoJd3(R-NqKWqC>) G7DuRLnL0 S! ˊdE3IM_ؑY5=UK!6=$$giɎWxV$AH6~NǣE1ݏo荒mh2ef;OѬ#i6il7wUY!kV5N) Gv\ .CgΈqd4/x&SГt,~OFNKJηp?܊{3lvdf {Cw;o;rd(/(eNiqMf 68&b\Yb-9߾ͷO"A0cs7]fEFdc..|xsqL7}sJ;_@0YטӍGtA;#4_~0UmJ( Gx<]hz%4\8kJ }磻6 `3(|]T\O`9A226} g$tbp5 ~d|p <.Wz{kV9Y~ =oƣ>`^o6I! B.Nuyw <9p*irf&pΰѴFe[g vZxg/)Iclr֒Y8 fKPg%󟽙^M~f*-tc x"ýv^2#YJߝ~w,x>x{x;/?>?|?rGx2I=|ksbw[ZwO[&'~6o(w>Z֨ݨp{U]z<њ6Cv! hjip~v/'AHg0ȕ1"c; %Q1(,F}F׼&; h})I!{+]?jlfܾBPKc(V!c2\s73 /y8Tf~'_0q)ToY J^f9f0_y3çρՃŋ! ՘LGx#\Iɰ:_?]6Q>kTKѫHL#!8*Z0iE\:Jq¸9hi2KSC7yÌLz3Id|A2bWB!L$z,&J[Q0!x* \!t6!3lEX#! !:WY.߀JXRpp%HQX(ꃉFP Lq0ɴ7I6 K;kh!a#4B!rORfϡ! HJ*2L,0a\hM9<3346WF#̱^{U^#Q@:B1Q( pmJrLr6Tˡ TE32bʹaV+Y2X25 94'ό@$`h$3~+ħ Cd5@ELI;2C$f21V]T.eTčue6wˌ=!& &-$$ a簼BߘV1[\|5xTr'U9&#pY3|?wC}\!cY5ڠ(M\ʜ0,4EƒdF?/baOqGˈV"\F5ss7]\y{{`1XnOZ.NGVg7U5 Gc 'VsU$cAE?GD?JhFV?Od<ѵTըK~'0B=3$;KQP>:w h bv2S pGuN0ve'vEؐ91;oS:"{&e)f/bϏ$WxE{ۘϟۘ("=ܠDەoxA}:b2h26g7NyY+2FqN)jM_xG#cjoWDjGh17fa%\16;Ex9|ͽ%^tŬx͛V%o.&Bh=tڤ@4NiAhPхZ;\=*5D=|A: 9Skȁ7o/dT댏=]eX.>ƞ9TUd'^[q+mE=SmBqmdc@]uzZ:Gs:^ʩnKbNĝӭdU\g9]>ݺ=LF8Gb5|u{4J,ti\գ]>0hT(Qi61˰=K}!׌1ٳ-ڛO((mc HM ArQ;BZ|6)EP)lTRuMlR Q#+A9aXFewjkXtak6*k5-d6)H<#aNӞjSwWPNUcV4*Y0|%*J>Rjb '!/U>.^r^g Ҋb /Ayn$ΩuxElTaIrSU\l;aTL~cuyfF\m'"I2ZL'ߖ85X2;d+l hf B;X4O')\Q/9X0pI3'@#Fŀ)2ތ$LC 2¸[YuJzZlT o o!?Jp,K}_G9@U5xN*ccsaq\NsN/MB/2mbF%L!Zb lj-\ kC,c$20duV}^inZal`ip`\~V΁3 ;7Ou7c B3?\-eAKC{<&tZ `Ӗƅ)Y8W`4 En'n|[]g ]yAˎ25!C4 S0}vY7I 6C*褎QƺuwC{WޚuKf4ֺՁ$LQ-u :cn;Hlͺ%3k@ hZT+̤hͿRG k&^^ B n6B/R+OY4Ԩi)3 .+Iy߻Ts:jT;gzY 9J8Xq`^g9ΐ#RCf?@~   /JAJ:FgE!ڞ3 Ձ$L${֍R܋cT@'u2֭Ct;Jf4ֺՁ$L {sߺ:ߡuKA tR(cݺEbvx-XVB^8DS0ILk3թi(cdJH Z%8t[h+Sxa1"YsB3 wsOȘXe"GI'rj[nU{TNHu1l=B֨aT1ʬu׿[_yM dϺq"NAT 0qRaμ)&3!/Ibu^& *+qR)"]$3bXB^8D0Uy?z{: "zOMfxGK(('Q&wDP)k&'5 C4 S~:@u⨢MH;SLqb]d.^K#3'5m{浐Pv,J1*Q8"q*CaLx3i qa_K8`X0 hy'HV!׻TzOMAqIݬ*[[Du C4Sޯ${ 6vMATQr:@gݒ΅!/IuMI/2}Se1X.wwhu!/1 egΉw<> &ӇOC]Ճ&|!5vf^r/!sEwK"3ٓq9 OrXk+8y}VE!6_cfBPM?IY B`F0H(6C(@hHi+Y,I}~yFCqBcL):A,>td8x.sx:U`|9M$/d86,,#&Dq~J#=lx_r|FN^f)'9<ӈJ`ui0~~ȇ8.r' `IBѦֽy(>#. #3wA̹%a`4s7M/@_d "ٯwё~7 [yfמ\jō`2Rz7GAX0]_ºl> y?\I.< {3["3yiϧ=TqU!֕x-7){˻yͭR;cȞ&5XqBi&t cjǛľup] dM@s<qD-vڭ_TF 0^- \ׄ;v &8y]Q:VH9.޹هtbr468k -lk Ш))hn0g`]~>ͫ?ă`z=#6ҝ;p`b+TFiM>RV-kU{Kh#grYLRt.!SD )Y*Qd:fMǼ{HJzSǴ<P,`*cgL"P%8K 6Yh6'Ez"6[X yx`el|l8feshU 5kMfkw)]xY:c+!]X7|K 3tӊwV\zxa8Ne&D&,r7 ^b3FE.Hg<*oxN U.bDtRԜN|.#=ctZw)ne8WFmְBޚ0/;35ڽ=\Qx I6g3`mBR\h9v~ޛv0>qY/{ARhu:*x փ vy=- ځPAxĄk'ʥu<*yØH#M" 58IJ&b!MSR- ׄ1^jџ,bª! on9.`8mft(&y)fI:b"rUF5Ihbak\cfw׈7Y(.-6j#iqCQV%jcu(_ti9p|^HnÇ?SUo;=M83ݍR=i 1W8!U∡0 2a2c,B+HET+a,/jfY:C 8qy:i7O0:SlLLBc#H)"FD()9:aҠ;I0LC.cvZp0iK, E񸢳CQe'XXZaH,1# i7""(Ji"_шS$ii^FW@Մj'B  *8ZQLXX"G-c\0L*tXJZ2S5~Z H>hU4XTsx^"5)X0AkE0Scۥ@'O*RE}6*S%~QYJOU UP9 Fr ŊoOfZȏ~.bTBoݣ|e, gUFd͌}`Ǥ~sir륯ôrwt/gYX4^yW$Q.rs{ / ZHz?93 ""@4_8އ/x0қyZ Cy7TB~o<T$BD$~$xΕ)Ø*?Z)DRF(OLMND~˺VՖo-@lƶxISU5QQ.MBtgTa.@vVVn]O88{2Օ2+'/6g c%q\t#tu\ݞ1y/9ōg"؂d\v0xL Ìl ,2=7\YPL esgl{zY,C89$}uHp_gɅuCUeV'>ƧV A$i3+`F԰*5Px\>x.rqlsd^m κC*.xΫR,CPUs]_ @bB|VV|UBwIfu'<k`ԑ|4T\p^q3DUCn}6{uybE mĝw)aT y b(ҲYs^V+׏yK6*WbDmPy}MKs4x&Lќ^y0"w=\-Ŵi 1ק܉ؓA|"5s}vC},Yˈ!ٙGm~tQ_vk63cwT(Z^cfwhvE[b+8k,DkaEv&8ꜸBD:C<;C{]Yz@bv n/\[6@(H X'yu#f ]gRuiD&q5t+hkS8Z N6ѢT+\w);ugO7>vQQBrNxsF4{y6H=v7`/k5>F #D7XͱB57 SD+q'xźj׍/x*V+ZEdӢISBBSgB6pMr rG9KqRD6Xװ%CgF7<ܥNbŃhPPտb&Ax2uez]`Ij=ż3Xpf JCsO҄SZKǒ+2N:kE8aD%NWg#L#bktV+#3 uW[6pW:{ntM,D5Uk Omv7Пc*޾.)XuqN~ t7@q@RB+xxSs9>Ut0D]nw:^fgjA'!oQSC2E NɳGLCH*+{vUw }a)zvlOBőjyg#&ۯ3y6{RX+C\ #:uz1[rZc$ pVͫ?=/GL6\{́oNb Ż,Bo7X(r1 a J"Lu(,6"PKsN rKȘ0268Ŕk-X.!'!%#RZ"f(%cHE07ԓ3zֹƤajCБq61IB0:LEZ^4BtՇoq!ż1ϼG=2?fJ .?P9&=_K%CN6Q([_?<Z?P#B$Q6;,Rq`=?@30pΧ3xa<oה'Ex<.6]$Nx};2IdIg}2qj 1b?P*Q*CkT5EW1)0`b+LeH A9n׃|]Z֪iH0&(2$š"0 A! U6U#A&1IDlIG4 )d{뭻q+7f.tS9uH c$zӌ 47.V1XK*L=rUZqd0'_6<15M W PJ3RtPN ԽB$+Mo8"9vgW0[qC9JzsƔv8Ͱ\QT{y;?3`q3XU#j Dw  XQeΚ' i^yz; N̯i%;^yy聛nBT;BGJ(X蠆Ge Lscll#iwϑ:8P{9UCdG:c, O?5P0ٜU |l8{ڕs$RYRX,{/fSK:,¿䐓I:Qݵ]I1eeyGzER FAX,l-V Kg-UBjSČ!šCeD,H,VW1j}b֡dI7LB59c`2"juirk% i ʉWşJGf(&Rzi1ґ`ׅ:Jߟu!,PxjZ%.v])'čW$ߕf|SB i̩~LF:+rػFr$W,i> ff]yA2ɒvl-u̠S$<]Ö8!ey(V<͇?L hvmDKVU}\X% 8\0W^׏zPKωp˨Мo/p,3jl.zU6e1tMjW/ E]U? b2IR%$%)FH7SVj`~z $ ۚQJ֑2Hֈ)I'ŪOhSva!R2Y4*뭙kyЅ7V]u˸h#S91zrړf gʸWXBOϗГ>kjצ?Sz%(E0{֢u'Towb#0-ktkL%m_Ni΅s6^g4,!6;Q\P୭f KP/9LGKM;VѪ;3y}d7Cw2#awKR'>鈎%s(4H5ƓXh`|Ḣ.OImNj9>s: Y ]Etw2\3i/L҇w3;>dJ!zaVwp-AҺo UwCű\bu{ɱŧ;Gq6buJ;fNH:Z l:LI`D]NC2G6טé;(,W]b)6]XBӕ(c,Jj'yصsW䦷D[L!}+]JBtzkgص&\_eǿ?x֞CŻJ0% ɩN hi)sxaм %_RPi|{x]Ra繹@ƯĵU Y":]{)bwLJO~psf/{\OJh 5U^X Ra R'el(g[YASdzCn>$ =~8,"mijdVs+dz쁸:ש&TVF<gUܸF{e;im:ېkJj%Z ic*AUt5WqRt~0 ?nO+sm\SҋڵQrԻ(*WOr8+fi m1ѵU6۰1 Kէa6Ůa{=jx,&g y3fO?3_M}>4l3βK Q c1$Pp8p5J{N]9Rޜ*-Q7)Ͽo'HdIeDp*۱SJ.U1#0-x~ntl=b.q,V `qV Ik2%PB5es s"JwDet sZN+0$*=:[pK{Dy#רG1Kw"SP_C%Ms0TyZZm&sp$bro˃W+88YsXd68{dc5 R/eג#z2 챭Qlݳ]`i՟qg_l?`{u7y?hIY>I{! :{mͬzg%/!©PB2Jh& I8}ZTpRIުMgN<_(*Zj+$EXhN[B圃 ru~;Z DӎJHFଁr]KN"LS%[?6T*a!p \>`pƔdNz 6ӓ) Uʳ U Q"+i4]$AcӀK/lCcP;A!#z힎V*tVa;x{Waڗ{H}Ļ/(Dj(J$uG h tTcCd}Z[->/ D8q幐@>Q\PZWЀsS1-ǔ3x[ `Xhڑf0|Ukvq̊9"wތAg<> &ӇaV5_ũo.^-,Jyr%qħͦy'"^$/. t0ƭnu}o c#˱99-aٚ9cD;S=21'_%'W#pCqDv3Lm"G&9ܲUM9#k|N?o.Of6(H=n.ZzV~sj9:飨 B!wT*&8 V)BȨ:CRILoPF1 QV`H*+<g y ƣ/D1 6|1S6($B=eiPDmRSE0:2`|Y-X3haT4lcĻ}ksΙgD7 )F!Jiec m3 "S؄L9Fq<(]iMڭY!pAV-_/w5H`$LܯDLH@> ba*\ *e!wD8*6ub]F4aSOE%e˰%SAW(Mj'FyImyð`*W -})ׯ52J2;ؤp妤r1IGsL_߉: גdK~}L~SăfR5l8Ƙ?[])NMw$!^ B)}RSߜ)UDgo Pe zEG,8?+$vq`Q`H@[AP()Uy:)gUֆ5UJ8*WC1]w(_ d}L@LiM^/D%H3r4WedJ,Ŋ4,l( -E"2G1s YF*.gw#?38d ||re7M|)5^ IA+]|J)ȵ+C09&U>Gs5zq:hO^w"um8՛X ?ŚЗ6K3_ 3L-O7m)nMqoMQf"y9L`@uY.J$H3 -׆B;{SP"\2mniTx)v*^h@v/JSx_f-1RInƺr 6*΄X!bE\aFsy4ͣ5L7nd, f1hC:k?5:{<.s~O,u̎nJ& ෋|p}uD?VR7wo׃|ds0XD^͟2]N}xE=Sx$dvdYJYjT 3YE[u+^c`9 Bgp0Ȥ &5e307[/~xH,ZXqHQ3O5?IsFq<۲3H K|HIX(4]I#@%&6S<[έrL39.wX=Dk FyKoVx]M-nM8ww1vB#F~|.2WZOĚVH± SA@/1SBaL\+Z84er+̝DfcL(VSZl^* M%oJq* Ԉ5+).C 29OrAX[E+8)&a0$kH4-֥63"䚧|DF? -0[PX Қ+SnkX~4})2Z.}3"1at T}3۞ dk]?/~ \n'Vv4Sseͦ;|ӌh.$b"OYͬ[FŶT Iƚ_-{f7=3ِTx8d oaދ\d2Ϝty͂FnvxGp5޴ Z?D9YUnt 2V2'a@D-b{x9eTeF678aK#EX#k1\;Y|.8Ŵf!yf:׵ >ʄ5i08[D P W8w9p/X/cĆsD$&(U\w^t1u65<ʌϚPJ k O1GhAPu $eXU S:F(0S"Y f'Z)R*!jARעTTjҧܥ*cƤMOT'3JEFEM(M d)0h6<1+RB0XV"h|gIhTqÒ)wAr,usDX egQLϢVNz">òZmpZR I9Ağ\Wxw5 $&7?{>r1:X _F{Ǜ٧C@T?{fX5[7'Jkowwe:>힩@՟Gued<| 4LJ긱m֊3Z,1=vXnna7>}o7ۀz@^۝WD ?_vITd0JJ3ămx}LK*۱0N}rTsT-SMHj{Ppt#[-Xb.;ۭuv6RHoV}9w`T{|E)g_Ѕh'C:-"~m&D<4Feyg ҨCv1Yj8/fXޛMhMl{'puvTY]se&G$#TJt ]9`(S1ifSX%no<1^({;tzQv0fUAKc3P29dl9X,39F(ӛ$j7O8E^3}z#f9֌I]* V;P]["P> |?|XxyՌǵL64CXիb/Xo/V}GL*2b·c=1bfN[\"@z.xKVLˉ{\λ_nC93~eS.XfHq̸bme!\жLHOˋEXajq{a'{J+VsQlse,#((Ŕ+4Z&NgNs`V>Vj^GR{] |2d<89cB83汦둜<76t-gÊW菫|*Djn9bS| F~57՘/փC'E\jCϚMh)%UFF g<[Zkt@TJ<! (Fx ót“"Jjy4RzQ@TtRL9+$@z!1|g?#Vߩ,v;i5VaH ˥(fL0K5 aƊwXLDKYE 9պh0 *[YSyo[UZ:-#ZNlߨEsU:Ip2vwgq1q˕Y5sM.-l؂6+lcGt\ !r~jb  q]^DK|rxD09n¯ό}L߫cv1WvEm l_/WY.c, +cSu _]LGK{?*?tas'W{C6 ߿mf 1)g1Ϗ[ d$8sbeX*z++_)(0E^ 6i5]S2X78%KV3YC~R Iot|t̋U3ͥ|]ʟ%vK$. !W{͗u59qI*" DKwjKW?)nJ'=\W"J?rDpSJrnݭZ6wO05JC-J θnrekţ3XFTX&MñxVѭ5@{8%6XTړݽ$8 L_T{ɂbjE@mVyf]  i Cc*7qL%v:6Y>nW[mq4 q=Hmé18Lg+^5')[rk^RS\ D58_ n@h@-%)fblLgW)Uv2~ vp2ϖYcs~a᳝=LݺM[LFyo? LfHڎk0w_f~x}XBy+ga4bBo?slNlW.@{Wh-=vcu`sG\D)_aZ+UC^woM8O$x1I~X,".p?oV'BGah{|^X.3bhDr%XVڑZF}001r'8VJ+=S sL.vV~@UBljۦɓ_F|{U4sO`9>~囕hD&N$քwCRTkJޏN|H"dL ySrEL.P@K#Xjp 0c9e#rRNh#ᨀc{c9xȺG~"eTQe٣1*+1W1Jj "=}+ı’ 9lvJڰD=y],$8.Vңhz'x (x4LuΘ ZpƆ(Dש:W`,D6f1 =&O `$8 ܿ4 x<EKi9䔖ӛ4ոOz85eMF')-8*8#=XDěܘP ŋqv=iBBP]hrbg,Iٵ\pSҦBX') oAu]1FV>UYz{6hP0#+~5'Md$j^p 9n!4d 1.IAL0H'7ӕoN/V9kyyW/ ~OF s/ߗD7ʇp~d-.&/N>rc[_n=;aP)Ihe"nhex,YX;'i\IQϵ%64H1QbZz}`ajWt"nmVShd:VzԆ6<:X?-QYupU]xycr.D9GoZJ{B;8J8yvASE, .ƃ|KjN]㯬hcXdZiŘ18HtV3ơ-ūᎈn(EM.8qM9›xñ$L_"%Â|1AX# .F$dRSXcb^k=EU:VA( N@![ꉠ\Q">2G9&ࠢnNPVz}T܁Y6FOgFDr鱖.9ZβD՘K6*1,n^|=rNe!1ΣAL\?LRۨΔS:dtٌTQPAK8iV/ariwfS)ښQGON$0Q,c7kф`!0V2Tu^^ {0 Y3Q=`].[L'z 1]!Yv*0mc,wZm/k&WdN&=K^=7N[]uCfa3 awyߖ)UE]3ƨ!j%yQwiu'@)XdG'䲫"}"iC=4gS*q&]26X {R+Y.Ϧ$4Re: ={yJj`3Zc*>j9oCrAI?f6-J.yEŧ6Bl` GiHP^WT%n }k6C>E"b4gD.B|c @T<3ˌ K3 Y#mpJ2jႸ ӱ 9˼L*E2L06sjfJCo c%Ж݈[qKZέɳ@c%!:S$p ǹzیQɭ1sǂ|jl#]=ؐ5P_\^G!xY=XD%C.MJQ&:aLQf99[4MGq, ,&<DB-O_~6?ǍZ<ŀK8AAm-і8 I,49KsVwuba`ffWw֪sRߖyaB!MԃKd”,学 1K%JʚIէT>uaCjfRݖ.mRݖ.mեΰR"[Sa>,$Sc>A12QCm+ O{`j+};ӎ @TFlF^=N*v*)yɒp`_̞#?7&S5Afr[.8'H4! gf(3(j &~@u(1kزHf0״`nk9-h:PhY$g(V^݂%!|t'pV[xtA=gA;#BP܁{=榼& y{Gvf~ݼȸ`Xn0cFURʉ bu/m@xŝ_a;)12 u>V3q ͘ /PʰXHHka]!_V)~IaH 4젽)߃bL=u\`|[Vq;w{iQu~QbeDqSa Ѳ 4T@U1f>rV'g *JY⑓:HD$H"P*G7/<r *bY)΍̦d 0U:9sj甡@/,?DfBp * lk}'\X?C-pn[Jʷ(ƀ{go+:a\6ؕ>LByyj L-i*>wYWtb;QU9qlDM[U]7+<̏r/SݔϟP!.99NyBIĩ;&NrIQwBdi#- A>>ŃS\$bP9UsGEd^$c,:K ڊ8B9 ^=NT}Y𖖠KTHkg y !,-&L&PaJSfRaFTXawXjB29fx ,[{}۲#~ev]kZʳrΎt= ;qu 7*2ϯ4]X>/1 8Wf0G-ru)vh>}lAQOob`.V-dH6W_W>/1{!fYZ=fe9WڲJC[QWo<$O z;';{Y~YçK]Ȟ8ׅ6-9-|zs= 4V#Վ>Iݧ:'PQjlNn3z?dD}D˄{~=4Kww{: !&b`}3ӛ,gɧ'%ȣ ɑ9$'O`-'\ky;٪BCh=#wR:*t.{Glf>Ë, _6pۘݕsV-Ǘ=d%Cm]~vt۱{Q82cu3iUC<{2!U~| ͫ )#wwG< ߋQ8sD1/w_V{Yp9&jA m/K[ Ę#l"yl!K`o뛕x~A+:M;uu(s(GE)i9OC ֱG/T;.ӫZR_>`ぅvOS }Ġ$bCKNn7~ԛC̕_#q)7Ms=fa9/~tOqo;܅7Sd1cʋ ͬ~<+8:IY>1ot躦$g[H hqxpXvG&IQ81Jj5ԢΦ=<[JrMLKKF(Iuh'Gd¨.OrytKô8O[#MxZxr1F~A /K_#4'5FI(b1G"ѱVK +BN)rHRQRG|z)ݔ1D6V㉢ɉj2E\)Jŋr9q.H2+w%@oH/nJ,WeJǫ U97`o@lus*Fp#y)$fM zU6HYz+}nP,67wlHWH&eMA3,wimIxTOv,S;~^i~8Kvicp~y/h>i[_b=ĸtg3[C=_>_j]~(gվ[gw Ɏ\+/ݯ=K\1PX"/0,jL WC@,HnA (]}ׯ`'ŸRc\w~\\I~)) qե=A!VhaϼZ_oH*E.7$_oHXߐhz!E{nHXߐbz!tiK  If9xG -.u-4΢ɢ68IpBΠb}CӋ )Rq!r}C.7Tr}^~z drϫo?(n;kgDl 3p_  ` fFn~oqNoY x k+y(,SfA#SPګR * $ bads:'c" fN@3CO;-.3au T~EG0K~ % Kߴ_}IXR -BYS0ᬓ泆TjT(* oØO o^ok>W5>qvAqugsesɰ:ZdOIjcS+Da<Q#uIQt؃ ~ täAɜJjr▥`>x~:qyY)SNmjЖ'9uԏ8mgRA*i tQ!9UG@OQ39p3(e""C򐋫,54Ox!(be.ҤI[MF iD_m1^e3o;}%yGғb99iҔu+r8g`g$g&[*Б#&9'׹2Qwv"‘Ύ%&}Hx:w&bv JYv,|bLjJ~eΑa)N1j>]K}vI0E@_-crNdONDz/[}ƶY{5(F[Ʃ;8gGRL/}ʺ=s3AhbL[oTOjQcxRɵfM[ʆ]˅5O:WYٙ!%Ϛ`^ZkŵϚԲXg U'Y~-/Zz쏛:?WxyynxUl*]_Kzɕ-X/n]<ؕuY2W*hOՊ\_|,D u}S߅|LύoQrSM3cfdlJ[) )eZ g$hTQb:cnO+iQg-tvkC.ESxJ5;wr-IvbISڝ^->R!΢I<tQb:cn4n "[r,SMwh7F*1{nİNu[t;Dj68Y44|PIR1a1h#@Ә-iNڭ pM⩦[Q1nDN[@Sn]b[ r,S~cɾYW:|^?l9pw2 o~}uʆ-z~[n nfjg{0SǨ{ITg9dvnC.ESxJ`5,&Ĺa*Q=qZg"cc}8Y4|6eL)su\S K,oS5e9NYmJD0̚15eMqJzloa֚){jʞjUt3>3mtʴ2mڔE)ݽt2v%AHȚջSdmKI,;4E.!<%Yux{]Bd x/&)QlM6%R4ȚCSd"kSM-zI"kSdUIPkYeg)6Eڔ]dM2a)v 5)FYLs:E֦Z_dSBYkS`gI{8+L3M}d81bLf,&%5YM5Ivۆm}YOfmeU;WZWMЄ^eM#fQ7QYӫ y[Yk+k{NU1Oj+kDeM5 i[Yk+kNg)meU8sĥYk+kyA/q[Yk+k Vj'8ׄMT֔@'8WmeU:DV}jƌ^eM#V־k2$k•w!_v~a40u>]v‡0.:wI u~5WY?$s֛.yp8`ʋA Vp?EYdD` aWT)12&R#%0 7y<'*%$qM_GPp4`IdD #Eoxuw2ʀf?->!٭ǥtKk赚gIͳYRkˮ'6JRP S`!1Gr9̙$;PMxkY}Ui,7*E)E/UjL&"$0"'G -lj0% &U H|> &-1igO/]v0/s8u@:ӄЌȔQQ].1za<|V5W="S0xGŤY(G7"Ƶq!" "cVca.!?󯀄J EFIcpOMB;#k`c3@^jҲ! >}&X$1 ;#8MC3\4A)D" ɒ04A1I(q$#Ffu* ! 5 Rn2P1p\\`(X QRR*g T B_ yJypH2hXBS|HA! 2DIMS8=H"gg8.Bfe*ֿ2w7j2Θt)n]Wu|^ _HC vJY,Mg)5Cg2/ –sѽ,~C?S`˔C3nYr7yY wP6gM 7(+_]\~D{Mxu;ɾĚ[gB3F#2 D4졠-κ[a"Q5U^gEmWm/^K ‡8o$s<!ψNd#W?@KMϋ 33co' ZDu(4ckXdJ)8|LaCjbpp\亩  Bd:%>.Kؿln(̙ˋF@J@ .JL'}]|. k9.W|2)Sj:`>p+0˵}a n=9dP 3{.~"p[TBt]t;߽5e |y '}`RD}(  GQ <3 Vix@tkԑ:!wGQ AYH.[d!0e(kAUd<<"/,ZB+:cf 0EC݈ȕpɩ21uiXeE`#C&dYU6d!_(JɅE{ e6t ?u׻.6$$+/,KǗ+=_Žsqq0.wJR\t7:tXW"Zӫ=8͟l` LS<2\ `ψ;#'ڰHs+̿"L ,w aLNB ,.,9"߼̵w@F+~+9ͧ`σ"7)fW6y˞1QJᛗs&hH#"Ciug]ͪݩ:o{ 0K=`%KgB?{FfV\ֽg$Ec>ےn\]H֣DP.z`_5#)=M:Yץ1BʒRT}/BJ36\t/6CjSװsى8W9f=6uIfkIc +[N룢f aqiD6Y;i*+sl<fОB}o|apAkF 3M@qK7_R53$m%*?VЧUV˴C5w$2'vHxC´fBݙuHC1%el`Y(W{\;8.j& ,>MyuKk(\n:9F7g)(UӚOV2y;EdqEVNJ խ,6߾2 DžZUd 䟫m-K<ő^6aTM`a檺G-(i9TpLF2Ƒό/e'm^OBӤ$=$,ghjw`=xO՞zȖM'Ag}(>z<.֗:.'.%-AN03[Kz$iQ30ΞkU_1e2MN#T+..j١^0Xg)L.F0\, 8$<kV;Dl͉03G9AqK Ik8ARPĮYUǁB-5VNKٝm`j'5f}W+Y؂J 0Q\=cp5 QJ{AqK`^$H@/N$7O\|]tMJJT688EQ;l1ւ) (G0 x#61 (#索A)|Iv"4_6VsRr}6ùf3=֊mBrH~\ϓ$(?A|DC{YaxCb_.k->:!8%"}HnnCb;G4'Qcj }AEOݑҕ!fc9Ă k&z$=ߛLj1jٗ>U6LȎCP.IpzHoaƏ΂|4jF-;½I|ģ gco xxL@ AR-hh%.isºAۙO3>6 q`R<5u%GL﯃ 498A/'B!Yf萩EƨB+D -f䬤ok aGNZ>=OR+N"i3|rbR {2!$Gj|里&HTRjRFřC-O W6ZxFz`nәܗ[ o3)% 89ꤡDI!1)Ds PE!pፈzjюMаZGC4Z8"m"* @bU'q@m'r`jI/q0K(esJ;y߹\[)$O[n%L{b"1wٵ8 # BGPXci^%GGoIn&*q^ !))표_W@%UOozcGA5|6GC]ch1& ;M@Jq|f)ϮcN;߅ñ ঳ZD.lBj=v( aOgÅl%>b,%+rodVdi>@˜ұĒg9$:Cl +1u'׊%@/|21+׌DeUgTh][6+,JEU{SJ\[*=h4Ǝ7~ԍ(+CQn4 Qql.FL `t2cA>$1ņ_N=`~Fr /Q2b-)x}\7ޡVsjR(c#uޢjXY[OdaE+#Qfӏc#}|aqBTG3u["ay>Uö]Q+u4̯{HR0x.nL1з~ eR'&K~~3ұs~:L~|!7|sr_lr1忇b{UQg~+3q'W_A_%&/]%o.\y_K W|Wh?2[uDu2bא,dV:q%<tC%ȑL9:j"eطɓqy:BFdh۝\1:W~|&q*}oA _#C&_aLIb%1) n5fk}+UQ8la_?E{}!އFH9kbo˼!!tH=04 <lIQHq1rw3FHTYi_"S_ESu趥ƈ nK=g8*qlKu['OCeZ?x>J5] 9 ;}t결A ;Ys"D{KPwB!,xG;gLlQ]3%ZơwЉ,) Z/"¼pAʩ !Bn -/KR D]) Ӡ\!uǤ+̂aiC+Ysb:K%}.rq54vf?Nc\XM.^0i1{þ^"klaQ hhZ!<a6M1V.JWy *TxNd:%HDt%ODnTmX zHQ⛧]%g"4ez8( /ђۗ"F~ `NE5놆q[n9us֗ш HTdɂt+i$*Q.S */Յ]*).Ajx{ %n)M&ek{`:V/4#)M}hâsh?CLq43goEACPNjI4]'aY1@UXp;q_0}9Ͽ,arZ[H F5{y<ⷎHE_V Ul}>"1EqC6Ǹū׳waNmW~ˬQ97Gޒ쁂hteXɑ5wg!*PaiwyubC;L`"锆N5 a|>%;TSwyX맧,[)\`i$1Sҏmx&ATz}d$L 5~aAq{SR`#"VʠH6#RL)/)(G'XĔK&%P8!:%xe!}Y< u0\@LX&3:Jaڱ d=;Wi_؛^C_2~bѝ׾B0fHdQVhmJ4B,bɭ L7!)R .Q$gdKo{f:}oufK(LIPا<# e2m 6J$"hBeJi=g#(I!QunO%b a+vnirxp~'omkli' A@쐫gn!H4! }`-Û &PIn, nl&>&o5.6Ϟv?pokYn-y1 ֻ|zJ8`#${ n{,4/tyifD_ b2zm!W-?M2K@S 9oAcY|n:SRd7`$#x0xkFJZC4ꏵq4[ڔk&c֑L!b%07 Dž+ ԩ{%e`4~UIZmDLv_O$A.Ux-/l3܀BHqT,n%h[C@fsh?=LGbW 塰C4.lL- vW>^|m-Fb*՛qѰPT`PZEPJ0x$uH`2@hHk#N-K*&Q 2,|L |OΌ !4ŜrmMEȩR$`U ,,gogL3 ~`&%&0hcyJdx&)NÞ)eH K gHJ. ̊Z.>{oAkDaEY"8S@-SWU߲ےKBbIHQB6sDRY7zJKL Os|k[I@ROg>̷94G(DžqsDU!~qrU8Cae,~Pm;6#U9nCے69#d ?c-2EcFBS`AcڤkJ@2FWG5iW +/*(\&t @1L'"eh6`MϑR"Ns`aT CvEH/E:iǕĪHo$mId&4dA_JQ%d_8*%"ceZ9,7|$ީ*e\;ᗔ壤=.c0B ƜeBu-AMΓX v%;0@B4&)BQB){IRՄli[ҦfH v2!)<\Á4rXAmy}n:nN9^yB{Lc1 Sq*8xu!5sg17OgIkAd75qHq(y%1"U*QtHt$/ hlսMuzֆSn<45f&HXz0nq>.xyMNL5ŴC%ɕ,xfJVUi|}*?.>xv7M,^qnFe,$t^l~Rt9`nUpW/WMp_G߅^#q[;|9Yku;U|Y$yk+kwzYnHsb^ =B-^f(_IԫU)J*IfiHMl7'TT|n90<ݷF>seoR5`%WxLΤ^;8 K-RN[l$>MM&ZdSxO21#tdv (Ix 12S)LƜapW` PJi*4sL!Ṟ$ф$h7"'D/҇z #3:#cŭ! ŕAi 9ϩwC.El vDZۜgff-zk*L6_]킼7Aw[}v0ˆoׯÇ!p:#y4y'|3p8wnMXn4f(pC9܉ +R}90%G%0ץQ_-VAjT/qQ\TQr*'JN]!PQN(r?[nMoO{Ɵx;Gfe+WD;*hO^[;ÛٛO]zHϖWWW)DžV)G n:Dɻ#%Y6<&6XºI z!%vQ!9|2r8JO$$eUᎧ/a3 =Anށ# cʘ{ruᕋscՕk_m﵋Kn})ke-Q< mلG}/f0sCfsrY͝_L4wFqaqjz/mVpcLs ^͑6" G4ϥ˅)'ËUV$J} &P;lW.T٭T1 ,nn"5H@Q!2Ys;y4QT;T֢^!7AW!{ V^{ה@`RNEDj5Xj|xeOE Wg6?j΋uoG-& H|{*%8Y{lvۛ=P=6a -nChXYAۻ_2l,U@XKuXUҒqM)}ljSF7) e?ZBJ9S9ךC[ e^"jeVJKx ͙R^Lv4.:.lqO[⏳YYmqs0OA*ѳ9 [:PR5Q Ec_JGgo1xO?#/o>9?Ṱ8F_\ pa΄z:ޱP*cE↭;LJn ӷX 0jFT7;31kEt6֝嚭U#,d᭕mKcD{1@R:/މj {MĠo*A h0h@a RqpIpYJ#8#ly՛zj!J%Wu|D֛EOy{ ;%KP~D·8Z G醭Cf<psx/9KANPy|`8l$RZ"Ԫ<]qPKTR,(y)%*M-|eC@!8MF|48m]d CnJlv*4@.8j `--3$LKfڈ*r0M_duɂߛ Og'TDfCN@(sf_˿?cI:ݡ͚Qԯ3KFɔ=Ul}MwvPIvm>8RsWCzB=S2tVX[iRw"Irp; 8>9#5cr5!62~*nO6\IOBF[5xpa(K] u& $3J"™@ҳ]HG yxdcؖ.&;gm^.6)M 'cx:ݵ.bNCmZ씕^BߥkӲSʢ~?m\} p]& 8qHrx::@MҎ!$` iLqʺvGux4J(U4SI@зD@(ޟ7F}"pPQY (Fe,i(=J),6EYYln}3'U 1q6n 8IInD}:O+Aw\R 's Q P1u\u80h߾;\gWuv]gWuvt5;hJ]BS.ђ8ت [D6HJ긚%F?4utA\!֡V:5ʫ+)`F92/95BF7R@ܷp4DX RB4)C=TS$ߩ_ZuC.B%L=:Qd^5`\$mH`&x awD b""sGD$X[ J!5W1JJkJQA @Q`ESf8Iu[~ ,!5AhEzF CG )B o,5Be }szKB"0侃Lɥ$RŁ241˜qbND5Fe&59p'ItY 0:hK1 Z Uߨ^Ncހ ,'܀c!c.$bX2 )4f~D2ɱQ*S%#Ҕ֖`;̔vp"k3-pm\H65Q>aL?ռ&gx0|ѡ3x 'kN"#P"qJ9d9dCL9`N5kwPd Lc;Ҍӽ2X?ΒH|ځ{͎DSΌnf^5UH#;9?!LG;3N40ll[ BDoC$0J I3i~!s_;8bOaԴ1{Ї_j^?2zf Pw.q1#H {=A⺮GVwntFq`z۸ۗ?>_QO` sM Hzx݌;?J1ydrlux~T-` G rQO";O@$ A? ,F(yHǻ;{t*rjFΜSY9MhwVUO g hlh4o8[&Tc ;6yx-D>Ymn3#F}k_-Pֆq]b&N2s+=ǽ8{P@d"c>~wv[>:G{B'||6inrǑG{)ŸFg?_[t} ܟVR/w3 kXij/L5ARҬp))HEay0lTB> ~u{s% AmqK}N  4\wָkrSWhayKh `k ` G^݆'IU2'<%Lq3syIAPmlƆ<Ѧ?D $zdxn՟DiIDЇCKj0Z)}`HHzS%H!D5!̺cѧ>+b8!2*-΀N\٬DSFC(Y4g+Q_-ݡ?X#f)@XD[˼&3²Xf5x<ʰv̑[ӵDtj ϫˆȷ'AΈ!l!QC^Q?_X,nLW`~& z XG-GocE; A{i_;׷T1HL"f#vk,^ G`1_EfjFcJp$TM,GLgTƖږYsM,/:NxG8sd Y QKq,ow>EbE$~w8BЧ3%9` *Jɭ3e ˔2#%G9* Ly! |› b'mDŬ-vZYX f6o XOBce 78;Dr;. HOmGjisrMIŌC +ۖٓWa 3HBvO)mdq9Msڮ( Ĥ.' UcT7D?E[pbs0=bpL0%LrN55'9e_ [y.aa@tWS8)1R]tcTnզJ*q?I5~okĸ}r]ywDxgqQOYA6 *ϜjU>Ou6'[mI/ǷX?z֑]=͚s>(Bk~?~pk Eo7F~2`ݍ}8gަS m[(Xv`W>v MW0< ZK%+,8$G|8L fvv3ď${-nnI@حY/U9R:kmM\PѤgZ@ i<1OVbMeGC;zuCH2Y\O%O}jzY}V?;zV+Ɍ/N)ɹO*FCV 3 J 8`AZXmtkSMK}hO$QyAѥFi1NwZӶ֥o!t^H F$tY0 H9 idЩZh.ygvT#!KDW-c"/ƪ1k*}<`1"R0=s~HLYɬ7eGZ͎Yẵ!șzZoPEbLA'ɬg,392JmK)F=##{׀329CS>i1ιyQ;W tyIRlUF2`xn~[-kTR ^x)$% E @-ak%>xڛ֬c*0ò 6ԘCѨb:y\A!3R$GvKI@ VedqcN:+ f>A\ɃtZ0G@nr&2TOv2#I%bFFfHIOMBK2ʄP>9Hn[)' &=Rkw&5YA)aN=I>#5jsuFlƢ$P뭼 kn'F'`5m"7\u' VN֕]I͆$f*r&hYZ_aF&ͳB$a+iɳt"8(Sr+Ѳ^`jXrb[oH0u뭱DBc:A? ޼XT͸K]*>56:Fƭ3-Y ǜBLNRp4>;5n(B o~0tO{<ΕNTY`9뵑m`l֝G0%՚q2ߠVN2u/ 5嚦& Qˆ ٙL6' Be Tax&IFKn ,ԄкCg%|$\1! G@9>'s_cW$V/zrk0t]U ZT#)}G&IHjgS>I^DW=KɘIn`A K GDį^zj~0t _R,ʖ~:'F{}v"zrl,M1EH 6DIpL6zc@4 . 5K kGj=V*d3 3d bƭPLKB-JHGbhĽ)Z@ƸܖrUTFˁ%'YDf2}3ޑ{K&?(V#BY*+X#4fg8)Y0%κK=F֕]; ~qj ZЌV2ruwm ̬zR>޶dLZo(IVjo l6Em{l״)5!ASgIQ7Ϥ(XL}r_/>e [ ҵzJ7iꚎ5\r4S2'P}>ò#uu&\}LStP;<p>bŖh`KT2Hz:KRw r0H  xzf`&2W7'&g1}huEvAJyd :6ii`}\f'WWIpӰ߀ʱRFcIn~͇*,;y,xpH)JлkT\ 2R]24ϊK"ZM/ױlǫuoI7 |/<ڊG%K6f+H[ O@1Ȍ[-~苧s2GO=mSf}VR.)$qglߨ,9E9ƤBd9{1Ov4;CVNh9UZ+9xd \5ء8vXy%͑F!P4YY{673v"xN 20e5NDQ6jA.ǏioSeF/[#PH`F))E#)JӲw  t{Xjr,V2U.$ /Dj9,zc6iFw&laޝM>5u`ɠ > E\p>\Zn\T)Lp~< fvڻζu4 !U}GH>f+H M˞V岇q]];eX}L]g#C4\fgLxezEebwG`~Nw k[C;̎3h2bE |MC8&ۃ% >`{: lo d{]r o?>ʑ"eN[¬z?Gc0GߝzwvyTh_?'zs{rDP1[b{DIZ_M* s$l>y8u{L=VЦ{L 吝1 1B?ոUу}`y8Q/y+~?`ԥ^?:YC+U!iÅSXґ1=l ARR~%Z@k_k,rrq2C"5qq:__ sX!d >ٟ&&ߡ ^/齭` ~>~rf<}WxnGd㔜E:OrBӻƲ.Rߔ'f'L.&FFYÍwQRQ➕#͐?~NTg c2n5$G5[1SK֙vPmPꒋ5gP+jsD5H>cPCa-_)+Ta;c`]m",}8D)2n3d6Մc߿ݐ+ƢxTJS319O(ZW7x(]ݠf_Oɢ듂BUXdÒޔt1cr _y,@NB9fWͤ\ q/u ޮPg ?6^$gwV(cH'K#o lyN/\;%}ݏ?nn]^WHЭE \Õ{4D?j\m#{ٞ0ɌVqw@&2s+ÿsqOrxVk Ē9wm>(C 1J̑hybEatھI{Lpb2+RW5VkpQ_}9ۄh՞NF?^H<h%G/&xlu}vs/=Qؕ^$8!pW"^%%apl .J"(F_i܅tr{h"04V hv-0 +CɈ4i{+6Ւ[5M[g=̝_+&,]FfJσAF.^Y_U1PŦ }<f!Z3ZܱRQ4\:/ⷎjg_d}f_k/(fH:y+=RE2Hvqԃ#MѠ !1|#p.`Y' nk[V!znEHX/rSl:j}`R"egƼRG Mm$u`M C^8Y^L>Z{'&*IS]0^fp5 㔬7>DJ.=2 0[/ekK]^!V! V+()8t{H&A"ts86!VGO2`l2'ipssO!OQذRXWc+o,qVUWtȢŜ%ݻud#^ ,HS3W6J)wwyef PBcdwik(%:8 %:49--*OC: 3jeҎOcL}0h8]PڣPd ӻs}ٻn$W,vGxk`v6dy!ۚm,w.%%:>,\ $n +WUFW,|i Z*J,$)Q80shk/19.n聗C ,v+`m:z.3{o֡UH>Tԛ_z'SDߜOm1|@=$AAc^&~iCO'='`E9c+Y]TdDHÕQLPC 7q`[Jt$DMELZO1WHTTp~hv].{4/7Wcb8Q",Fv94"hU\* RRQ!~,U{n(ޮh_֭Ld*)xW M*)BhQleY1Є>+@W+dkzV=Zb: <:cS\%P\sfҦ$oA;<WOpEx%k%[͵ɇlKW$M׋~iïiيl'~2?̬51,Cw4ٵVb'핎%U(H/1Jpw1~rBho*VJ?SBw&t+3<Ms\0I͸O=?z<9 $)A{]`!?  #'삫:/k5 Du^?hXT9Uߎcћ$I3]T (W0rqu稶޷դOQRw 8li.mMwޢ>Tzeݙg>3?=81 D1:p_K(6>78 "90ġષ+#r0WKo \ ^Ƀq<$=Sz-D-])Z^:W$nv4+A/ui$*Wa,UHVxk6VW*X<Ysӭ?h?k?/3i[Z!F sK!8(w폣O=j{cMd7i?k9:qg˟V1߯>$a?blqc׋^/<+{kݼZ3mTS~:{$udxkaP]ޮ~壡T-cjc]vy2~[$8ÆM)s ޑrͲ=;M#iD#Ż{yLa8_u:һua!DlT#]5׈-W1;Fw!?׷waݺWn ,YTrBw[$_nW~s0:Xzqs9ܹn>8 /W7zy^gfpu)^MKR{*sj,8svnDTB`D8u9kϰyÂ"Bu KbZ< FY&C{W07攥` O}N-sr57;SJaU;I3:jB^"*JV9J98`Rr=(#0@RU-l`P=C6t3"ոtl$|j!:Z^P 9`f IDX,pZTHpKqEenƸLizN T0ta T ͞$>U}W~Cg)EL g1)x+G-ĘOpWB0ug _;4Y-onz;Ǣ[xʹ=㽉_&~Ln/0u)_V*`H[(9v1A:iYs8|?[4Vny`yX^3ìK%}dvF6G7)웫&gDVZ$tlSN8O|H;kQiΕ?|hD44%(u [r0Q&n<)B hwaEk? ]X|wm,ӻkv<8ih*02܋6pԇnDDPls.+|kYvw C=H{SF0ܭP:TKW'a J.%E@!]q:*071s\U̻W) Ϗ̸ ;"J[V#&Zrg1TeHVt%vn#S-*AXͭ%T%^Ƶ#^)j*38B)9g1%<\'MgiIkdkiF2R24!S!7b4ڎMxaoWG^c_)vg7\T(%OMt{\WOf^x5瑐v7a.˛w؛wR(y@Ofuq7Çtwa\{q}[Ui%}o?WKݫR2+C]ɨ+(D;]Z.t7o&/שy!iۏ\Y ,6CV0X*4i;Is= u*gdJJU˒xc͆ƚn]/tHT!RX;|y1k d7c߳!Ah/PsP:LE&pRү:X!4a>j:C,*.pUpHq_`^~\a#E7{='$uO\0v\$YkSP6&%]ifuBN(Pbɇv "TL1 Q_D%%_4FRƨ"6"MmWupK6ςzo)v⇙&mt-ޝg?=iu?1jlm (đ ^{N~ rrSZб7yZ#e_1+DHz/W_::OovklWF;ZqWWͱCO'wgW?'oYOf1E }x t(ui:es؜+{) wfݨkVQH$ m Ɔ.?BtSBQqm3H\$3{g;=R$@5;Mv]!Qڵ͆>gbO]SID(,[?;1qxf]!g\ˁGX(KV'K>j7x#AL1Rr?xLr@mWnY6ԻM֚ruc:cx/yޭ y&eS.x7keGn:11Rۈ4 &yޭ y&ڰȌ+YcL2102],m ÐTd16{ Vv$XܮJ<҇L7惓Ȯg_|>+7CWM#sxLEz/Gz͹/LO"( 4A0twX{q$:%^b q8zzļ"f,`d0EDїhi.! SmgQ ᨠM> (FGTD9a:IŞ)qv|tkJ$)` cz٫,6FLTHd`S^CX*ϝI*Dz]ZH!y:tLk`6Y^K= }E2_!RWK۴5m6-3%(WY $)>y_9ko:y#N01t8*h~ I'B^fTvF); {(c$Ƌf?r GͯHKB^f'qz7u-W1;FwAUo_8һua!D[۔w/"R_7RTK(gF_,iP0UUX7?_;]ԚF+&K.f {5B0-4x {~7_j{oogdGb^cuK''s@g2{ ݎI+3:(N{D(5 @d. xHj}ƂDP9$]v_/~BAe JUm [ykN 餒BnnOCI܅k|YŴ.I'YHuarNr"=Т!0faO)dde RNb;2`6B ѤR_gkn_^}5("uudc SE©lm=O!e_eϟϳ䐼F{񴔯x#.8khIC4ü5ԁN!k*Í繾2O !viR>79BT!X"ںط#Fܥ'ࢪܱꮲ?HA Ek/C^8[S3 i8evo8er jolo>v1s}x-Y=) ʷϤs* "+d!J81CŁ1[V@~DǡYeBsN">pW4@LKtOf61s8p'WHi ABx-G(ABi ę Ƹ]<^C9fGPܩ8EbDG3x%Qv;һL LI'wf{[5cfpa`P #`2%FB,Q(Tb c~=P7fGͿx#T׋g{6WG.L7Mm& B-ŲlKwghɦآH :lږHǜs>=(&sZ6]-Žעq`bPTJI l,IҲ BI>5NJɜwyv,NYBr`4=E.X'ɰ+S 5!Mai5V6ej˶Wv9JJyãE9CD78WWiIzdEt]ꮫWJ+Xg̃稴V!&q3UXr^LaU]HKNXmU% t>IRx(NNtt4%I|ݹ%@IJ-d`,A%Lt4^DbN1}^&(" VkA&9‰w+T:1蚘@M W",pӏmhN{&9PyA)ĈEUƝ z)=c, .oU ߪa 津leWT"S}*(?B[O8$BeXHAAZt7_do*sXӾɔ'Rخbtϝ<*e;eH:E2$ J|c\wdBUfIwo3EOUϼWA$R_^ڹ]!܏&WZ9 B\\ d!nFds*w='<|Ym/p_x)‰DOX3+%~Sǯ2 pVZkvpY~ֶRIclͨB<>|%Kv,uJMoVO=U7։.VN+GNR'8y`ZY:*WԜ{/qF  c[[rFpTv*]Ц@8jrl}smv[5ª*̽gɻwr,_}&J.Yr6]_*կr&J#C5&+@=8Y~巵ӯ$,Ku4?<6k:9W*udPR "d{ۅJqlX3#xB{3v|0 6|4Ç>*eEmOրP/y e%h暡W}m6o > UTWMDu F J4_ R{"{X{UƽP%hSz0m(O3oWnLP-+%`9ݺAtc[2'DrXIRJ9S5KSSOT?T1SԪN\[uPJǀCw8T wU=Eɥmm}TkWFei`6ڍ&~*  ;濫 +&IsZ&/o^~/|&>`7%/œ/0ge a1*3 `)EΓw\2\% yasi*J\`B֙☂ƈc9(pQWĶR2x ~K 3K GM 7a/vH_ e1l AGao^[XVćhbcJ䀉qKBwZTHQo0&Nv7s6ya:BK [h,אH -]zxұnoOB$]ڐxv;5Ns %>beWjq[~ B l`"X3(HnrM-M خa$&Dyl2 &Z 'zdR_RYƛHƛ@%Y $hk'N*"%`ӭ>>(c5 !i^8L~/Z\iq{*U@MogC"@c1^ NpA%)*K%<+`٦ YP͑p>Ws"XPq_Y:H:^JyF* AIߎcD\AD\AW,V`G,Ƚ˵ c"\!H <O=9,AY/*yӁy9H:IY k };bA}kMM3冁8a!0pF ,&Y.S]}o:P}놃sF5W)ļylhptNKN)aUH>[H(n?=*i. ]| W,78^y4=}1wM0W/'~pegwrmk{low޽|ο4)ӣd4)puv^>{o{Ȫ{޽ۗ_,ް۳_[ܓ,xf?/6f?frwGf*8v1`L}N`foZ(~\0(~X Rwo~ _8%Яeףټ 1x__ 5o={f/ϋyb/'Ȍ_Z l;0GOv?{۶?ZgnAv[M eVT=8F;dJvlRl9@*|~Ùnc0Ix~5Lk8g= sχ>XOϷ; =v t67 éIoo+Xѻ~>3'xb4v+v/8%Q72j؟ 0`wG@g\v 5O0d=>%3oz3 )wpPWn1;/ 3LBrSpetBGs)N/W@07Qj 4ܗl N-6ˠY@Enu3pBfAjfcwg0)(La;SqqK{~Z_S__(DYESVeƖN AT0EecGDF.\ |݇1Yc)a<~1vk]cccy3),ȉ".iL$ Tj \Aؒ܊ǂW"$1f3մ+zHRӨ<Jqx,{.p8F-/ 0F[onC1,hCVS9 Dmmj 6DZE\.T?L%L\KX4QilJ9t+$KZ ZcŗW*??\;msy;sBLI&Rj-rI(jE(1fl`b_C7Qت/V/SFPښ1#'h{r''h{rv`wMHF, So,SK1Oh(5#. &SER%DjWkޠV&YNBPC%RIDi .V` Pbd뺝:r )9:Pr  EX% dCIV0CҦa'%ks\/#O[BcN3-8AKIa ΂;a.Q æ f8ibM [T.7a>0Vhp0 yd:3*xG  UMHw5EĐCO y>1+" \U| onwmPHX-'U6bU*E<҄k4Nxˡ `i X$` #QdSq ͫ&&F*oKJ<5XSc<5V!_Lxd3aUyE,`&f$I2+"^b sŀ4I1KcsBL$1HTT/Dij#\V1Z*0B y+)P|DUt#_.)"ȓ*"O򤊅B֊4 Q1Mw.ExQFGyeQ."ުU $0VQXoB"bqM)&s*V.ȸbClSolvIxa*92H)mBPS@~B 8Naj]y Sb4 _:BTW)eJo2䛦s"aLᎹ>Ks5uz1{ u$s&4:tGAw[ޓ: :&Mc:Y_澺>ib%HخLr)d( CK/NnIQСJ:Yϊu( eى] b}N)~,_8WPچ[G?.?\Wܞr-gb(a`0s}(,k(+ ON|/rN{Zg הhX Hx 'Bj!&_zTѓtzc7a;})J:={v>"9&'_W5z J\63I՘Q5B{<OF/Ҍ"џW ;gyP=K{jNmAcoq%ozCNFk((OQܷ/Yܮspw4x3@&TsUkƂ:iw8$RFi;%۹Ҥ;7>yc7tY4hf>TxB]0&${=of#( wI zX-l-Q|l*04(ٸe']m]ig^=6 B'8a*Ԥy,I?I0Fs>8t]h;w=,}$FgLVkԫ4e0ஶ_P$MWi4l~\\3MBu?I9!\憝4p'f ao<<+vN{9<Iq~?2c(pDy,K#{qA`|cEѵ֣#U.;WD#o$z40d5t#l;q- ka^(d|~Vav za*{\aA<C`mг=a̯[UkQRFipV܊HJmGܕz:Q1L(Eqߞ\|YifJǢk3 oIƷd5Vf| abƗ]R!SkKQbIrK(׃*kC4ٕ݊Jncp!''nk*ÍَR|<[WvxH8oT֍8N !:v_ƅvǿsFUbF" -dyf%K2V8a,q acAhTE| izɵ Uk PJpESSg.>z4_F|>m͇'B]yjQ烐BY=;="kn4zXbft ̥H [ͣ>Xgsߚs8  / {swrxoQw ~t{?z=_Opn7~=i[{ /z5]f?0i:|Ӂoo:e2fSľ5.poOSx8SaCh_]ӋJ pУ(,|5Tj繐}n5ĥS69_;c|%9?Їph|z~ }O (Fxy2] cƦ=2~ p9w7ϔ:+ |zzʍ[Ϧ?Be֍Tj]Fo͟qyiS~f%yu|GWG/~;Qwu Q)xd:Nopbpx yu;3̟ 4p*s{n$fVCxó^ AQ <S[(l4T{|]ܷԻ߇ǧL~`hݠLS'ry4OduΞ/ä^:2 7:Ko)=ߖWjFpɑO汋y~+ J -*b ,PfAwnܻjT\և;x=wM.a22X3)mQjHb37ҩ4\&Eq㱐DvNaS kǯuX[uX[uXBYFڔ'H|G1il )%猣cR'DR1ږa9]IWqP}/+TJm,3}L>gz:_%E? R ڈ75iClA{N͐pxG>5þOש0//l|^~ ))|dն]]_bup.*?vby{?ߟygb^|&2<?M)K[Vb[[\/,ώ>t/Ŷù}v{}ewp]ҽ(>_=K㿛.Fj~C1֫NICez_W_G-ڻJkpwog_X!m~XoZdr6}&}W _}%#BNRtAԙ> 6W -Ck5ҵ䴲z["0%+@^'Ž\C.In06Hl'wQ|vv}v_rzL\fqn;yusuh NOo`.*Q;2eJAh`&E kg!&՘6F,~13o0~yL r# `xfViYeEيZqR"6Tn@ ㄉ(cRfa.;J;-3fC bUZ)GZ-5LByLSڄy%!_=C*_5gmLF}_Eقjxݮ-ie'KK=4EZ1h @D/MD~! Z`aa|@] f %Vr iIdvЌsOG:- kl\ .◇ IR\p K)-I !m;4>6 %0nzIs# 3^3D2SKg6&#ju23Sk"0|BNEeIrѬ8N`[|O8+go2݄9?I֏oîHΞ٥~ǯ5{-)Hgxsvkzc4D{|NIJI%O6 b +p85B_N}c)a" ޴!mdJLU"J8\JY:p.#)Igs옲B$mOJыExМwOa&:T;bq_^u $F>u&;_ic(prZP\%Ya&+O@~S=](קi>k %t8~n+sl8)ڃ`|S'Ԣ|< xBALX|UU}+7V3D2r3]ϸ"X'_ f7Qθ=D-DnyP>83CiBi$>6Dz k m%VҤ+--Iڞ!\Eo/co!tac/ӡä "+)QtQ5V -eo#`dB 4u&&*MEPAS v',bq3Q3r%Sś K}1%xJO?<̶CdBFZ.Hh Q"x frѭ(9A<8:E2{EZr\rI]k扔spݭpJJDh*&:/|86ϡA)Qu p$IibS_|x#&6hD&D^ 6۵msX\IM)`srw.[l4ʏQDթZsT?()8w)%lVK8',O4ZՓEhH4z3;bahtV́[[!v03w/Ka =|bҰlH0xV !h%@%&lKҜKRjF;°j ±.TA/|xN#`2Е_j>Z]|ArvfO)%qJ+KFjt9bIɐZ"#,Q*)9.^0S=]*:Q~ZC6%P}AH'9} LryڶN: @10k `]} u# <dv,$t7tk=:X9$g;ɼaw6("!m.HDpJQO壑N!G!מBR܂ r"`=U^a,"tt u"D] KCA䤪~#iA7b\LT#)bD5bBIjcjq(9>`8D $K;;4VۇHքu?~AP-IhPj}w 8ڐ:@̈́r<`Xѧ>BގGjƴdd0΍ nb4 k}8Dn{q4R<ڦ^$CC/DG$$l $T>_#c2@Hx{ wYm4LV`ؘ{ncI!zPzDXLF[.FX;A9=%jv)n2)/EKBYFV2Gڋ5h<Ӻwܴ2N!*~%X8%B YЇݻF]G%fg妧o`ӱ20-)%i]/{ꣅrN8{Ӏ\ e߇UYY18MRӺm){m\mϘOxIhn;s>a4dq|BJ+@݋d戗y>Ib45'* "1>6$F^}c z}0kZӑF}嫯^&:)MizǯH;Kߛ`s_j:HyX ےߕbtg6kRH;>RL)ԑ'<{\bF|.YE۵7]o'ږ6ϻO ڏG?_ȭ AWoϷf}qU.?ӬY(v1ԙ\׬Eٜ(m;ˋ/2‹gO>&grewHS9'n w?KjW?8ً*܌:7fH0{;-vAd*1L:%/ /ڱq MZOfIn`ݺXE5SYg$D>QhY*,\1)]LXY̑``zƱ_) 0; 4;/(H$rue߇]eٖl"e)RT%.;\0Hf/1d4O3ō3aTR),OBv̺1bCʬ=RP1+H*"i&W<3~d-]ɾӇ %%7p$7;`3)`yGݣTc^4X!~G {M<\;ڇ4VI{I bN;:cUwV0Qq$8bQ (;!E1Vi{WZ 7v8f7 $̂ì[8J 4Kplڊg:2gvbVq&|_xb0$9Q }, <ӈBpĄ1I2'2?0<^ˉPJL<]dc\ @Ĉz( ʥc 8\q0Q˲3wRe # ȲX;`EXW$u!-\ڳHF ~$nލ$ NsW_xGLc[lI!4d[ou ѷ0Zc=_ZX M~C k҈Vֶyu$$]`Ks <`eakuSUD&gؕcYwkߪ+]}|ճ& 7$ZMdh*\`qRAZ\353d-l&8,MP"yv/+/"!.9a2 uV/{&GW>H,)l- jVBvR"i %LTI X`!˱p0y \$ Z5k jɐZ* C۠SmKb;g@TFJ Ui{.T)3'* ?z3Z7Z;VDv O!?>#:r(ML%>$7Tyh#d1bsCf+>0RI)HG\'OZ)M DB} pdH"}~ oBwX4{ɪ?727&kދ@ /b@HO 5Ȉ~Gw̿).eҔ!\=G@+79rpj8~ENGr>S$3˕-V_؝cwN99vWc')q]TcR@h&\܍>H* "zV>>7ڡipG3[%F}*@{eT949%>GރCԋјT Ayo"y?3c1lc7c2hՇ*1xx|rA zQ^:;*l%=)O׺tlwP#(.W9//qAk6v׷^CstyO:uPWr{vG*E`.zsi76qsi76)ϥA$6豝8uӽqpZ=+(.+hAZh]!(*ϟBv0 u=TN3)XͩR 8ZX!is8BFa+ƙd@4Ξ|(@ƭ )ԜKT!f rzMi1 94$UȤ'd,c:4{x-BhUWԭ7!VR:%H>E =co=RބVx@-h1k@.bZKD==dVDIf%ܜO%&`y~,H C*$8JeF"Lsϲ}ë|K+ 琂2 β%˯)ы%5S0|x=w'OXT>zP,`]Dxˀ n:{CjL ){=$>HD;/]?ij>+01[7aw )ĺ8V^t.OE ?cL.s \vTH:5"z!@,֟< a⮑[ zi@GT*ZYxE1K`0BձLf{=򘋈D^ġdTLçOaz:":_JFJY6>wKc.%qBLH"ڧD)Y 蠎*֯&Ƙ|`/%$xT$STX).0KfbR")D\DžG "FRc' H$LoD3 wKjɕj"plD-M}/T>s,a_)Ldy~ )^ww_9@(u;J&C.ojY|KfYG ϳ{q f(lLJ@( ďn[Ln.Ȕr/X'0h>,q_ej1Sˊz8$0vǝJ'Y`~/m[k#mimvknxDRV[v-'inJ#;v"ǒ$PGy!iGIJ7[Ճw^K`oBP1>v(Tr2kG5ê51!V15k,ٷ*E,o~L-uZ-҂DO-+Ї6J]hxE|`Z},uyFi4>fבiT DNZ}$U?RG/jl9W,L6r on$Z?01,Cry8.ۊ%b3soZ A>R/ hxc%K>Rc\%b6? @6A6Àsdo=@UÞ$XyZ#,B&7u*WQ@-5nuWopҀ\btk'vJ`@asT-l[@}kĿPz*VCĒ;>@͘MӂUR2nw Q mjxid՛J*YQƃUd!qFCZOlYO&}.ׇZ)Br=,V(T*Gv@orĹ=[PA<'I$--XA[qߕ19yv;Mv41ݭ@GbU8C-uAU6[@\u!{RrKc#Mx#HkH7tѿ4%[8EؗrI-aKfumG~)JiG8!} =0cb4OeCnSPP_}>fix[ |) 1;2WT[ "Q8ro_۱g6Og+O<{u#Fu A">bh! A/ܾpmݣA#+1&7HHHHA(|_3vK%1 !`۝1⃛N%Lљ[0 \p:rqS˨M-$;U ' :1l"Ψ-q1hDXDk\LאHFT2A|%RnI+9.-lSvxݺ_ѬZ\W"Dn)FQL}xOS.%@WuY\9 rNu}UV`t1ُôKAjږԻ|\r7:eQثY~z{g.JzXQQ XH1oiWp4a]ui$b )tCO;\RJ `xg&I=F^sDIxW[GǘD c=KfSHwoqDa?_x0(ⶣqYROt`}K=FKuu(cͷb3>!Rm$UWdVyF} [vZF \ ~ƅ Z\܊.slˊxĈ/=bTM\~)tI:rIsE`6GrhMg )LW'ny*V!骐BnIڥk +ohSaR8 !aG# G CnAF7i:ߖUֶl3m!,)~XU2B3Ge<ls߻;w+jԃYJ8+iu1?)C֬ůI`&*)ơȹMwaZw76W S?+AH*v$h\n)\Go,uTSsQ@@g%~cyi )|qρ(Z*-v*+jT݊} ua`[ x_qphfۻ2) f>-g+`c L L)4)lIҽ:zb$<6QCth%:Rq(L d`$Q(,VB ^-qIjp(Ϥ Nr {ՇϦgKCREt^ǎFFC.睗A-$XHғNAۯc<^/En8& $` SUbS{: RP_WFƛ> (:R3G α`I&+wi2~n79胀!00/F'Y/OwLK(a>ΜN\Iw;l2 ړxpm3blxz *=KM7G^}g$x$%\H޴T3En-[נw fpM2+9J:?g0 Q23ݸ 3 ɟ\k/_>Nm7&h" 3 3 ׹Es82521p=91ᯌ59=FA{[^xO;ʠ`wfA "tfo{1V.|>uGddIg~2K(4~e/HNKWN KP:, J ̛wN~~?G`tt4SI`B>wjnYA,YClxg Մf 87ݐ{b&~ S m~%I/N"/t8ݎI+$XoꞫtA/Wofֽ ˢ-~pe0 BgBt+]0`$?[wzGx`5Q;@u~}3tdŻ/?!o`0d3]+^op>+8zVc2[ aPԹ=geॡc>t䘃P:tkwxΚN*b8|1FWߏAB ؓfM,ͷe.tp|v痮p`oו%P빙y$]"&XNqR=q"0D4kv /6(FZ~ }bE |uv EPl<؀0U>C>'4aB.R턃4s2لL@za/WLЃp|C< |..qߤ<hWJQvdz(MBotKWi.u $5(x5+ Î KIc |ؑk2n 8e`ʜQ8pD3h6caRx´| i8dg@y%9>T)?o瓨Ͽ<-ϯsrNy$~?VD6 9 Pre$RDi6a)"5%EpzQ}ų'`,"DX3;`\5Nڹﳫ.rGUㅗ|.@>i]"KV̅?prW09 c 9GD`A X d*RY(oZH9/~i"_vpc ٨""G,t,&x,BJ{tC(P[dAW*ܠk,&e&\45Ʈ_`K`.AG_w 7r.T2j ?;JłgtNOE;$&t6P#L⦕TX婷bܤ*jN߲rЉ2XvAI'ә&/by4l{}ѷ}s}#5'ոG̋4,U`qi]!_1js=0>:~}|;7Jyw Rb7nMy1%UΌ%bR qKwDpòM Q -dv|:fFnj!;@Ϯ/v|N7Y砸zP\^NZHp>PyOl7Mvh %w 8-YLAL*}e/6= L?I\-Y̍ᤰOP[O  wAFUQtrhU-ë%^ޢesMͶ3P7M]A=kaN*zx\^s4 D`P{_}f3]t\`|vU?4]i.!IF=. a,uU\hVʜAk}#b-37D3765ɍ1tn[~j$h[4vl_ "^pWjÉ5';8;| qGpUλPi(KzңHO`I)f&"6?U5\}ζ1jЯ _ VQ$hXiS+;g^tq~>IBã5t7k-'E`pRݿ0E-}^ xcyhWj^֫z5ZFըëƬ^֫z5K~*pqȄF{*BN&Pѓg\dTQ=JH˭G,}/+0UJ6=%T=Jк7u[RPn׃r׀!.ILgӘJWtN +.w=arF5NgUep}O$]UQcfI۳65OXGICoS׌^Ifo}&ܽ< H*\&Q-S_SIϑA/F(O "=n"d͛Ey L $B`-/DhbQ&8X P$eo Nky$ 4%|hLx Ia)r*^a$#,,8@ Pkjt y%2b3$9ĸO D +>3gIPL*!Ғ*j ZbÍB&[1>@x4 ̴">Tkn,0Б")Fq8 ql9>6Vqm}L/ROiIi;/.JS_&$SeOI"b><a2{mV_:6^PZx0sʝ*Ҁ{ HL rxjƚ<`&Djnۭ#A6A5ⶢ! ֟3AD'u5s0R"DGUSbN/*Z} yLbsrSmA%h!s=*bZ<{9J:e:dʦ0܃<ΰ`/g|B޾kB⯛$j&3 .b|R07s_񟔢 uH%ޝ a:f}3KSe@NRL/g`}f5as-` I;J\̱6yT4=#LVaI; =Wl--}Q:Ђƞ_Lߴl;;/4waA6Ex&E~O>c3asFx3ceYtr)njނwr!)i -a0 \H-h/!kG=ΪD>iIp[7I[wB~ -zuuui{Қ ]).a. EL'N&Oh׀ cNrxC#Ľ$("ZXJd9ం;k KGcizG/Bq !ɽ8?He0 ^ `SN0e9 BќZbj/I͜5M1:'GINQwQ=ޝVm!b AAKFFB[6c\)@/Khj`ˬۃD~[Q+&!8f$r8Bb[H̱Z eFxч|p22`L QZiC 1+bӽ8F%AQx<`^p 3G@(O48#Ac&`$pI$ΊswgC^ͧ ^B Oyqw][Tk9%1`%)#pJNYpDy"W2,5 Y<;߁=:h:?yv~t6_ b3DWrbrfvt?]P!iQghfoRWqќ]/JLjQg쫋ݪh.K ST/LrT̼蒹Uj" \W쐢V ꪄЮZFdD$BC [&EԦt.6x*u.”큤]x SRD(oR0Ҷne11C$,t?mTIe$"eLw_-q;7}E(?zOJ#`iʊQHPYըsL.:0qh9& OH.i!N~wO1 ip9d""~@(n?A&0ib Α[S2bnU %{SjX3=2TiB#{ݑf,xk눽 ۀvbJKLQ x(1t߇y^\[ I]%xs``zPn-D S-@f V\FE\EFz(ssb1dX!d%+jg̪Q`Bq'ׅacZn.|4p]@3ВU.RdZ]l DD? 6`ܭjT53 o`b<!cA{U0v:=26nr[pN ޱ{+*+( XHVHG^F8-JjԹ}{wlDG h+YB #Í!P&ę+J K=\ɍ^MF4څ;J+fy%^#[VILAdxiHK fR/Šz;4OQ%B^ܭj@TC.q No6N1TǯtHhhgcg2%bj3.T 5k(_k,9ո*WQ@@{s*CZ4w7I;BWs-]PpPF۠YդY<(n ȅ`"<7`/KJ\ ĀO8cf(RHleբȪu|Zߝ5W흪? "q[^V+S2AH5y<B5rTe y2 6h|J봝r3ߏP`zCq*٬AVZ#4jlURx!( *ǦIHQP8nd@7R>Er{K|'%6 [Fu w e9|XWPͥh)0{?lgdsMb\<ذI*D1,3x"n(חcZyH5G'GQgCFo7{ lxMx"ϒjLnx:b%drT>{2Yeƻ+ڑN1溔kwfd~$Ja/3͌*3QS'٢ͥ`Nrir݄п/2Gؿd8(f7ifI8FY()Z8|0Zq tC_·yQn©h#eSjZ]&h!0[)Ĩ8Ps,PFjtO7 piEr}GxWrI˝nXd&ުy_ ]p;vm$: Q:!hj+9]J$/C+xU#t>[|H(Xoi/ hoe:AݖY!n[ m,E]K8Vp({Ϟ׻O&Ul:mtQTcɥOik-l6@)NeRS0UG$s5+T,oaEy]U/4l#"D6ViaUbWBl)XݥW+X`K"'k쪘  oxų1@vr>4O1<'GӉ- Gr<<3##˝YnǓErM5}GEO0WקߩYVx:+]!z?_1i^L~BL0SDisV. ~}y_/$|sMBÓo㝬>-&% }Vʟ[D~<~M+`cهP׿="[-?ytC|O‹7~(./@`zh z g8E,O6ůM@F^ښ~b~bBq/MX2͇O73y78[Oҭ1墕WT Bd:4s_&Z#7~xrk"[w.we|YDcg'1d DB9\YTRtuAoo"7)wd~[00Ѯ}i T i"><39o:7oO?@[ >.z1I gcXjJ?Gy4&K>KFsȸжOիNg ]hw$\iҷ_#7~?+rEUb2y:=O>ø,g7Z|{ C2AyW[1( -5{C|yT-,vz21^ܗ:8ȃQm=[/Jz6"0 '׀[{ZϢ;1xw"Ud󻪳͟4.' m5J!7J A /c$ZgvqaR[t1J .#^j_^(]1WE|Bt .thS@æph4x|f캹TۗMy3-F9=y*JRƝcScGJYccZ9D ' <&X'N1%e{uFb~8ݧuk6qJRbNzYǫ ]u8QWku#ݿ9% 5*>eI CQ퉞&i#YVtJF7NIdI$t3ڊsJ>EZ'v?%4r`W)N~0JێTfRg7y\>AJhR3+W䬪0wXs$<;cZqYz*m]SnPM},? TE:q+$TWZVpi]X* sNy{\'BY8" >7oǿև@T8VYy\Z1Ӽz.<|;-˱Iw֠r`o"ehTP-Dl㻶'z7iki>nX48<.B(GUu 'ݬLpt>8]F4;O: %P aon^YIFR<'BXf_Lb|O+-szSILj (2xR=mrBH H^ۄwE-bl+dz `U@DD$UsLPQpaM%"m!< J"4-Kܒn& Mlۤ1gJ U ,!n s]'o턞=ais7zF+aajҀI@$ϱQ-i9X(8s  !z<&Z,kY!WD0K rt3B'ry5񴘳6/>޹ן]͠9%~}7`=W`G72da$jU8Dgnl mDW~C5 G~^_\+f@ $'2||G]5xRFx܃jV wD@L}k&a.fH&B]6Kصm-EaQ֘LrTg\\aVq.Vpf0 #\59 8ta0H(k z ٟDg](V:J0b5T\ D#<U:!)V+z獵έLMQfScfn׬QܷllӶwYAhkO2KF% E1#Ԝ&hvKwHmK|jͺ+1OI1xv|۷c<3wcXƏdڛȨSv%rqWu83ۮyw$q&}ۼ\nB^fٔ½_z7zr1H11x.c)W;"]{MnmX+7 Ғ˻OZa+] di= !5".UQW.--8Jsxͪ2 a J/XIL T-J!sfW)^;+u3Ul':miES.V׏VZ9<,I58vEIY~zl _统?b8TqI_'Y2R,sMDgh:EDbNJ&Jho+)[Md@No>%W\" s["(_0ͬņ1sjnKTP{=6|Oŝ~HNG϶2MהIX^S*8!֟q +J>(t63~pn'gd^i~:)7jszljO7ZO[Zez XoOƟzK")Em)Szq)_e9DmHmCUM'DqޓW'k:KN1[|. XRvB?C-tt@+Å`q+'ܷ.9iSw<ʫZ(iGu[QF) )Y\*).Ôtxӵj#ZIJ%xAHy[NPL iGbr$ԊkpU]HJ8%/}Ci z^hOUaT阐ptߚIcD{F5.U3yTqB?(I38zEsJT%#Re- #BntwSA\=xEpWXhE52@^sԳӯpʈ _QA:m_#| ڧωLIt.c|mpxf(e+F 4/'7CY0}ӚnI~ +dW*_O<.X,S͕,]k1)w.H} %?}1$bb0oG[jVg?~90]໛k5CUG`_2d’=>ʺiA:Q\r㘲p])jY3NCt'Hu!vr],"KP&pwĴs1s\Fu.zbb(jvo j: iM &(:0X@ڹ"0HY%ptiXRLy{뛢C#MH0o47r-A2 m4_=Υ~uN>Iȯo+߫~f4P3|}#~t1"]1joRzr+WOPW$zN}pƌޑe[z\\IF$FdYZ3IIIJ꾓] KM$'E/: ^.rZYo@"9T1 &OQbe.hg$l'zBmۛ)WW_M´$uV˾'|`t1 9դ3W &R?X%zkB݁m*}?y;S2Gv4Zcw7Ai O#|sw2Iڶ c|>8?;}o|>%]Nn}w?h}Sܕo'.):tBvKoO;<`mmTΡVr4N~382' 5PU*:/\*bR XT K2pKTN2clq Q8qE*d0<~ȼA90bb*$V%@_H[)r-0BPBlX%9(7'Kk"tz" @1FPJJ+(1:x0w͐%އ4|vRgBG\ ߸R=m2͎x;2.;ؕEM!"JX[ZeD\m86Eig ^9Қ9xYD!xd F92V`pRERz$1IHI83%"j))G*b\sU4a< DOFO ,1ELdr )9"8d;mЂ^+h% =j՜9JJ3% c`$4^DV@ ,ӥ| -4'V2*SY.h]4$ȶlsG U*,ͥ.1%d>[t`+VB V[6(i9’y',cì: >zޗ#>"TqվЀHAD` Y("q1,`HD3#\qC"*qac&y|p7#(&, A ˅(ٕgvRE۷ODɴN=ƊaUw=f9]YKM4˦}XnB[.);Fvtt(ۛw~TֆrͰ)L&~/OJ1aޱ{+Y~]t+wm=n?ރv l 3#Jd~̴n#j9gDŪ"YU˖?k=F+)f#eɕK֬ח Ϗglv.LgVŊͼ:gA|q0_Sa-y%'%rxyinLooIчſ̫#x}ޛW|2ono7I4gݑf5vw$%L 2He2c  49=B|e+7?y5-Tk *2>jQ ,Q:C gv5;}bW G*MÖLOڃ_l{KH.%%mĕC44s&RBl<ʒc]Q΀O;픘g#IŹc|SB}?%}YCQ$91,9k9:So>ŝ=(uc__wL螒>9!8*(l-0x t_ UͽB^1N7rp|gP]'zƬtÁ}l0%mU#fR*֗kz:`.uu,9DtQw)!Sq6`Ҋ /!D(1J[QQ8(42Ki PL-sFث-ۏ([!狢nY90ƅ2H+0O L`[څo Q xNMq׻቟>o_`bp?Pz u=+`;6 Zd" VLH\i-,F2s1sƧ!SwNMsḀ锽++{rYـ]bH?wݾuB h[tw];G)96/u7Jy\ŻpM̟bExP֗aY#U#@: w( U  3^Fu0A}׀Y90岈2:6Ϛm;`g;}>iF9n B%3Pgsy>(*:A+8,Qqp,Zˢc^dg9Yw`"fq#9 b]ya/7q@`p2V pR]sf#SS(j^lm~ vJѲj|& ~=*sf/F"=1aTєkYk65q:e܂߇{F,ert-zk> 3((H}ASQ7_-~%E nG~rh*KhQ`>I0'SFV]!@j %v)fQY9s0#5z]?b^ʭ_*^d9A܎{sX{vq>P&[\b 66>;U-Xc.^A #6CV\tmgӌ 浃,g(p"lI c&UKeYA1q5At48haBU'^#d-ZV?UJ ˂;+IW 4ZEztzv{ÏMR2H8Ƙh9^j:ϐ(IiT'G,@09yblJ7R"Jz1ZB`mO*a]N~ϗnh Ϋ`n;FpZ-|2b?LyU_9ߪy|(U{PQM=op:Sr:0>[VRȏ%N,ç_ւز@ow>`z&g._>0Ơ!ݩ,N N 9pcZf7S \Շ_&C8o[4\p_')ۏ?LGˎ~%,kՒaykdq؜}W!΀@BҵTWW{%Yi<%+d#}=uk4Q ?v&A?}>Bd`|Z6?Ȧǽsi5El`Lqzg3'#ey,3s <I1gu ї}NᠬuY5)8q̩3E:X{\jX;k:v4 baMAGFbV ƙV\Tq.QHcfQu}Z귏Un7d7dsg0X GF#`Z4°@BМX/ %$Af9Ǥ%xv"O3(`#ER9(XRG7f0ۖRꊈϯ|0THTHU?G.EIIUJE7Rb}oyi*R+OUPaI</B.8AzkDm;؝Q 5Sƭ6* hg!6A%9#S`W/Zjf^^5ò`Ln sJmY?w[Fy 'UepG {5JofF2Su+  ˱K&(!7HՂ0-V{/!?A$CbYl):Oa N.2YKej=Zl_ofŬ"R_ׯ֟_]|qmn_^m=bѫQ+I-H+W?#^,Կskf󿬮j~ ?N&>Mp@RXs-~-( Nq)`sMO Y;j@xD V4BSGEg;Ok9sz@q'5Vje:/Dh ;>`v:-Ec[ûSIxB.r +L rb64DsT5h0Gg`,qzq#]kp(kqzB5NO ոBb> ;ePN0w&*Ҏ j8%kҁ^C^ L!:8%{\O]ObK+A5FJ^gŤFJJmEX2Q:P|^ Y/jTrUSqZe]>,D('94,x-G!%ku֎qڑPq=kG]> BJ +L@ @{Ř +cR*0B:[>J*OtH>9N)88zӺ|(-A H2X+Hv H͉Fkĸ'PQ:P|~Md{5Ϝ1ڵg9"(xoA6[' FbdK6^mxGSxXQ?̾\x :9o1 {Qa)l'ENqu 6ȅ806N阼qkLu""+"7Jh[Z^ɸXP6' cߋ`cR`Ysa rіRZsBÉauV1"'wfybHKBm9XA򃂛Ҕ߱EXh(1/CBJ5I#0p ^ 5 0AN, ҧH)2% 7Eh)Z;x(aTbSUMU)6UcS[:D9{h0% L$e%[Vۅ_PJto=9pK`St:+).V/, Q+-'Cz&DiQ9h-X`a.(L IutF籛ғ%W`6 qYuQ"tPLɪ؀1q,`ʀ6rVрAlmaq`Ёq&JY8!Edc΢`L@liVP,ewY.ųH(TwV$" _Z%>Dk"E~c`qJ:Ș^1R\6TⲬt(-ДHKK Ac-&wd (Uݓhk̉ޓG# ӧ&K`WZ41EXV~^FVW+"hp/s%^ӌ@%[v`9`^b ޝN5& ,oBF⸿ď?%YqRcb,Z)Ase}8c.k-/XJո\n X5縿4W+4Wֽ3s4JsQRNr 44DSz ^a 4Bd`?x|֬r{h@&X /Ɯ8Eu`=Gk^bXiYh6|4ô7V͗+Ձ1֐Q,2sN¯Swspd5S4L',;Z0Y(=G6U8ݚˎ*,[.TTqml.[?8* __io;6V>y%@K.zs}P~+gWyȵC,UV{@3"{6ORǍ]+k"]`!ӕ^^$p$=iN!Q_Sp9P}gSR o4!Y-MNIֽ3s4=Zestٺw n4稀5]O<9_.D"j.]I|![> G S.k1 .1_뚁&ԕ'94;R >ec%.4%`yqu2SWB; -Tu&Z8ahSeDgS ^ eTU<]yޙJtǵG '})94f4P"_Lf~UݜN$2W]W }X~lv$m={|ZOx .3~ + E|qaWNi jz: X3 +P!Dq&ETÝ=,M5,QֹSQg%h]V]Bv=EX]v`*㍯1ʽG} PM' 1Ҙ)1$AtHP<|q{e3šbQ!ovaA3j0``8RBW%'UrD8)*.:Tg1 C2JܸRd ߭a,*Ɩت?7SXOny23y3Xn.V?xHd75x8^+3!o4< ~2&`} cj̕_PH%#lFiPFn &E-]T,oꮙ*7h&J;uAb-j^UJ.9 >S[74>ۮ[ih"ui8fl\HhaCUї=r1CELZVX@!#%Bp*E̺suՖ9"4#Í!`B&J-5'd1,5M^ hMٶʦ]c71$W8S`|&{Z ]=B 2̫Y.Ip"yBP{uڤyps7%~IkV7=d3]hFW@Hwfȓ'?QŲԅVXXG5Iec\T̺hyQJ)o?Y\VCrェ*I{8[s'obdwj oQnnUu6CcSK۪罫iQ$xfpkgt:;FUcp:v:m>kW\ 5W:[/ssB(rs@ PiZ+1WY aC$z}<}Dx%>&Rpu7ׅOn`QOy֢t%ǐ%%B#͗@B >}Ԁ6sݼ~@e|gbK#<]"P'FJar8 `P"u^b@+`,[#IEp X}A,!o{KBbD^0 m=-7>D@nz1"e"F-TB8[*0 @ \ m+K`,E=Y4M!ܿ+B}W+MdfHBALLV4#=6VHp z<#SSћwRs! F7)!Uw2nvg!6_ڙ/8/.]Z^~ao"vg>7oVCMxvߤ}M3E.ZJ[9u|^qng|٦6˽y$,xpDΖz6S#|SǚuӨ,к5 4u;;57ָ֭)I8n1OM-dI.Br'$ˠ !^:ģbTHY(Q MB)Rb9#\un: hqp rX gF)Ǣp'-k !jŐ 8-)h5j'(tJJH}(0?ܖ_ɚz]_\=hhU_Ȗ,^yݫggR;ri_9)2F;h=@*y9U*1?]G|!/["GSWq4NTct3aBpK_u,d0r# ԲiEcRVJ|G,2VKFM_w~!$S,?Ms]3]^aM,8;5)ak7-XyY\a800b߿$)`"Gj7 IpI |~H/ y~~<[P6t3#Xòugpu=`|.Jx9Sx{X%ߢ ȭ/797ˍY/oX+vXSg{lAt :v`:icnH&`c! ?1QϦ1J(Iwώu"jsZڭH3n67j!x+@tlޡXP4wXɯIޟq?ED)c>z6cuYjhrvu]W$m2hg'hW!ٜ0Y  Y0m\a(ATYdS enB*˙7[c 5zƶV&6qJkjɨ}-dnk8V!>ssb$Y $(9B#+e0H`´;-1ibֱ3ﰼD8c:||QoccAAľ0~sܒ#T|5TŰd#xRlƕ3Gg? ?E5X ZT ۋA'ӹ"~xw7s5ǚ \QzL2^^r HPWc)HU(0A9԰MbRP >_ >> P zmw:&ȪY5}e۪BTh*4T _Jd7K~[ ҏdޠAҕx? ,DMܴMܴU7m^ )R1 QSS1#1;=OdXO-RoЗ7w}_ Vf2M0}'P37Ӳ6T*4U7|L`UtQ$2kޖk9%la|xlQ€ .J<* A(ʥXD1)s-'1 X(9IJab&6LS PCRPO)`}08sF"ʔ 2MÐE&f z/LL IB f8,ذC)Fgzȑ_YfaLcw^fa𴅖%ultdVf1nYR*" "Es0&G\0Sl 2/M;y4 ʑ{ox&;Jy!d4)% `0>)>zt/0!GrwDohm<# qe2D9">3 9҆C$6EeJ$l魛ӽAq@5>u_ 7=]*vvƯ((N5Oru%˂[,\`H#K *%<~Y}8s e'4!%Li36@BI1>!hQM&2-1 CciA4B aI5LQ3-\Br4\:C8, \9,e x# J ]J7r"S d3Qf6@zlL$<*IɦJAHaj\ͻgqVø7PZYɯ;YA[A .&WBJI 'J+(⴦r]$xK*1/6uMN@'\ 7M_-TsC۞kK]w EҵE;A I_Jz-EM [^5Wvm{} i;3H\WM XIwiCPC`f&22mF*Q.lLEvĒ: W@O5¦4Hھ"l$Z\#K ]5a6 -l5~js]CG8,1H`% E), 7I)&DxLueMn&dg( \P,#rd:OA-A\"OOY<~ 0u~LdgTQ,"@ zNu$dȚ*iEZHIfq\?K\ F ~͇L9rJș,.T1:ͭEdd'`w8LNEhHH'A;Yf%g%5I\a7s;CV](]'0 PLCC@6p=eZkcɴ] 4&n$36X+kԥdVXG>*: gt׾-nJnC7/nkC2Q7˱qhB#J Нc-է^pۥ5{=DȽ`#.m] Hk!u{UJ5)m.dIzlf?^4y:E6^4KldYa"Y~c"6ZY5c*I։5;K[ Țg 8d6*3s&10֎pL`ٯ Ul03rϙ0 *o=ofgrmTsE1WGkAtc7/d+5~s9:1w7+G2jJWo;Bn۞ oa 6"^‚oVu)cR};Ѩ ʚMhfqnH;cG\B7h-1ZG*Ewv-Y t& YǵZ֚{|HG-!QzX1NrC/W7TxZiH^Fzff#)P FOSnlyهui `,5^B^x\O@+p8uelG54zQJ},ܭRJB v}l$#&`rkTt㯎уw:gb&`a۩tSH! `9S1&#q`ʬ։f'ggrؚܱ.R^Jkw:nB/s̉UP҂Fm"26Y-x&T|svGI[_*v :snFR$ \yn"xxr.D* @ޘ ٭h2VצD2% zʌ>`tB4cՉEKu_ge=dw+cҕ0F~4`:b߿{%P&ʆG̍T5NG8eT 0p%$[75CLQh%M1z$>IVNvV̰tx5Xa@~uKΉ5闉̚g~"XH[v<7ԁO,+UR@oaX}#ЧR?5zi?Nnܫ47SUZF\O +|!:R""@^7 M8Yo^@-7Pf Q.cX9: q0/Sv%KC0fȜJeW:gAK^p=͂1D>f2!* Rmz<{W#)4]gmSW V; BŴSE.p\RMk#/`%N{ Gy.j\8BUrGݫ#xUmSzajTẑs0[Q!/G{mYy46=0,@ZNy] ב9<=Nnk)ښgξuDJ k˛:mysdmysme9-l_|,8|RJw e儏 L6_Q029c;* ^wGcFyY p=Zš<&&&JH/<f @Yv ,gAc'cFkBlW9pW#:AηBݠ|TsRNʿh7?;:ue[Zb!:T'ZCr p)E^jfF>£Y+Ӛ]GrolS !v!BK2_Bˇ=@;+T[,׷&o~/seݑkqJ-Ywa[ArC~?ͫ4 uzŅKtr޿%ocvAv;BJnCH+%R2&WNqqp;9,urEzͻ}~^īi(;ww+aj:}0uD!7e.s?4e@Jlrӝ:<\]gPκlâod]#G a%}%;^/i"WFr؄ 0Ta.7)?df7tVK|B5 9 zC2 m9)ApPygee:29O`lѤ sSڍvd9O+?-‡R#3rI56 CG=rtH4x  <OpKQyO9mv* 9OU}IGʯa}I7 qWšd{)WP{Q#+95J{|%]u4vX*7q[fO/ Mi02;<;qЧP[JB*Ar~lu :!Bpuh1;&Y@ٰY@ݘ$ziud'̼i?aqg:]tqjp쑿pPxTp벧:*Z-CnO?sT88|vKW#=RLRb~*&\ɌN3Gn ̒`/8XxD{lv me{|J-"-RE׀ߊX80^

gDmn%p LQzgru+fN?,+iGg@ԚXKK4J̈́]lپRmEϦÉ&7L K/#ײaMJ%Ea |o_W -Wwb2gi3 ӫY÷ YlX EVm#_b+-(1 d=_:X\ $cuWX~ω_Dg߽o ld m ;(A6R7UR?}'P}[q76S6y3yGC ȻF'[2qh dKUΔfCI^=[!r䕷/3S(}& S>#ŘD ЇݳaTAKttrQYxm# N$\[̟MUK ׌V W[ylھ\4VhV;]lK5n*;(Lx|fԬް܆1Y٧g~ ԳJnL̈S)?@` &}p8ĆU:J\=0F i+<>xgj{FW␤E+.\7-]rΖINrWw%Yd\jc^z6C~;3;Dps 9`F +XP+#-g<'_,K狕{u4uCH$휥|[U"YRJa1IOhV 0dLU;k%^|^]K|T,eIj ^΍Kӷ%')S^|,Qbxk*35&aLcr0EZyЊ4]MO|a3h¸L9>Λ[>CrHq!zJ1 *g%" B $f~09Lվ*0u'emp;WFfcF?l+A)gTB ygNE1G q`K2Kcrcۇ; ,c}mm"9?6:ɴ%\$584ZJy)ʳ6*ZVУYO ȸ< ]c˲4(( hBd3 |rMݞق݈acܚK|Į)vٗ˧6J|+%a:=(Eѧ3;9FP+Utx0 Dݰ40uvwDr‽n!4ŌR㌚8_DLG⢻1OcD㡶t6 0.Ȁj G8`G {xבƸ7q :G7J NR0[H-yJyǡЈ~3@@µxUK}0 8k4Zj?bVY^Ȭ?b_.o+@ت7 ݊W݊W݊W݊Wo.I9ܸTjrT3\#!(F[hP0e(XjxޯQ~9yX&"_w0 _]k0 Zsv~',GN`}0/&k4]"-d}]Ga4t\w5/HeP+s! X>B U1,p%(k̽B!HbPb0|>oGB-7a7зX4B'_Dޯ"k! FZFdI¡24] U\C俓]pgR]m?ĩff}uwGz &Ւy'YWʼnDUDVk)TeTg i|HsI4˝ùBVaygy4yߟMc.\(^OkDVH0U.3,r%Ri,2 ׄ?g<)Q\R{ P5hrO_#A8ois{dFB# G *Gr QW YYru!PRJ竂#zJ`[ bkdNhJR3s :ˤU&uXR9QJWDj2qTe kE uG'_  & 0hh``^+9)"O &:ZIt8溲,v͂1oBv:dr-DaN!|37w_fՅuq,-6~?w醫'ٳbbw9lv(Bw7ͯk`dO{x2Bͩ]SCjy~w2x*bOF%R\[V}BensdiqxQX|>OSάG D#I0$']:\O9Z.jZ0@p Ӯ Wt2Ony.dpt'm):dH)ُ1 _%o#;'Tԇ-[ S,z-NpLH#$AK:9XXk0m~{ZQH8w{:6 h;¯uӯtozxOP9]F~mwa <.KsLixxкRT1yCAPo'AW60ZWqMn/miMxW=n=V-FU\PRڥ-oF7OP-IG,]?ݻg[-i=Z3ތn=ܬFh98Emn5O%8uNNaoKH'$Ӣə^L f̥FY*9ݤ0)čf9QRHL2'}R75/d[ϯn_V S4#(..3̦b>=77w_KM)'F[/Z`+^U}𫟦w;uA I_g|X>VuSUQݳۓWHK.,y!1߶տuUO/ .&w[t]Nj#w?}Q[0x^ ޡJqX^ʷ_Rgw@3zv8lyUq,ps iKE E#-x0dj%q2&,z-gES˹I݋5|<|?f'ky _~&w,q׫K/ŵ37\dSs3YkEO?``Fmpr'.U-+uo|Y%N4x:Do\᲼|V2s5GsUbr g3#/r]Y@o\ߺt­^% >/(,=dJqR+?[)DHM"˓7ozgY~JD14R'Æ>t~U&p1e~tw)(~h1]ԲCh15{atY%A؋"ncMAL>w0--"cRԲ*q#tqDt@tղkU:s$QS$JQ˞z]kTy8N: JݲVMI;tgt+]~[z0uTP$Eq:aܔcqs1蒟-~-B9呖kj`.C# gXNF$؆$aT#(א"8DQD|(ڤ$/ ^}=>{Ooh(Vmg\j'no8wǝrM5- o\M,S.& (XR3|n䋻ɝ-vGI/^<.`h2-v0D:Ep>wExL> [ QBqǀwC3~FID^_.|TrikjM\yUa_TVz(pE´ JH'_ySк-IH:, ^~/>gOS>/͹EN&Y̾j("=A.֝@FcZD26(GMH VT}TɎG blaB0:D;T9k3쉚č7*E[ĎՈ=;=enuf#2|5Jʠa [3ꈑ1"`EX6<czlW"钉b5gˋo/AlSmIcbwy6ϸ1Pm- .K_wcq,fbԄ6 DЄ1[8-:1tkGcoɹ}6/d@V\N즆REDNr֕⽝RVX(%[wm5niDlCo):ffdV IX.tQ&In;@l\mʃf^.B(KBUD3[1,ZgiD&Dav;#F1ĕ]Uj* hLWNj,%AvUf2]yeW~,ZuIZbIŨn+0ZOa0vniqUobZeG |˷vSM,0䞹բm]l%ߊ2Ij3$:{t'.Lq<$dSP&dY:mZ6yۘW&ca_Jw-Y9[Y]@LVPfTJb"KrWeK6ݕ68nBAf9~±2WVkX% mLlJD +K .>/$"$Ap%7wrw;_dӅ W/k/~6(2)-ғL)Erǂ4 g3)=ቹ3%]:vE_d<ċ//Pq+ Y4Z v,/N*_~A9uϗD-m,45D|Z&Z#B+"[4J,):.Y2BkMW;gmi~cRW4mى32H^jR{ef꣨% & YBvPn~]UbG1WS;P.%~J "b@aGk3d2-p͞"_3{+oR ii?3=ql7Ƙz$A#`w0}Y#p}46y^1Ym:KLLRT~v:cPpB%)s ٻƍdWrҷg$ vi-;ld~%_hYI [Uuuk] >6bNȖ5Pԡ!. B6H AA(4E$cxPEygz8F8?e1)0w #("N-RZ9}6#+ : Xud"ܽ(L(s@KExtc:/  ̻^gw hnC-k!i]ӓz-W蟑p, [7St?]3샊cJ95cJe*?I4{iJ+I4$ k7H#x7nlQFhuO6\0Is(`uO|d ܛH4˃UufbUJмƆ!qtt"O.8{Y!+%yZm7` Ǧ V66Cc 1+j|'ޙO7do]|OfY@*Pp?pe2ؕ8T C7$$su;nۚi'!GY%In-) vNɴS2}DȚQSdH|B W8v2La/%GiGN(jh!r=F1*rW颸)lC$J85J#bJv~rqul*׆kB[5L53쥰Z(c$mzghJ\eX %4+)IfQWRRYcgd}>e"@yT* :!lnwr4y1s YK%wYʽt՞C}>_xM)F޲^ͳHyֱFԓGxyɒGcgCc9 NĹ`6_hC[I啕}0e~P|JR)k WouVE:u>|\,ye^,A2@ IwoGD,~3f"Y^L\Ь+X* /4nv>ϒ v'XkP̹EK* 5AƤbRfDdZ&5氢wm߻fI Ǐ{$ e鮌ꃣFx)=2H5>є$wz!vw{(6mx d!FiyJ6䢉`$D7{e(|j]5Soxx=:V/~H]Eq|eM܅vpL~hXcf]Q21leγC5RRC(:Sz4/4C|v3m-/Z)x9Ht{!oԩY] ]G11s]axTY.idq1OBng`q삱_+1c&N[x-ݢQ#8u~_VVEQu1`;1+:c:pkרzLC2uH9Bڪ]AkbWZ}do7~ )t[D`vNjڇRP+M)F9X&j?Sdh'}D3-;2̶h6$ˤ$gxpAAue[sRS/n.ՀP(bAl!imó%ď%3Xn6 P%i^۱ߞoA6/iz旌@4QTeh]p#맋?ڿ'% Ϳ_L1~9b+ sj6KX\zLSHjS"^W@3!,cr Q.Txoe*x8&oksNB^o|Zb~q6C'l2=^xFUz?H7  l䗫?嚣Ź/,z7Y&Y_PTϐqҬ3yKDhM۴Д h'ܑʂ h[NisfHڄaA|vG\mb6vR$;;C 2Ea2`3SDVk=Cw:g(MZF陁YlM ~܅A5ܗ> eZtqr0얰L{JsJay.Zm' H;I⤑/$ 7_Ͻ *~Ԩ{\90ЪUëg 3Ӻ DdPtƨN[wYMU<雡ǐMMu~4p.Q%qMWDh{u/pˋ;(SwEP9V vuB [Y߼2;Ҽ(K\}O~8=yPٗʾ$T%/)V$T⻿{++UkNz=zFj,7f~d6psLܭ9a%KmQ 25&;| aMSU,W, 4bk"*gD Bl`3sg' a`TKʭ~,E\dlc /zX!id#>2NP[. 48͌TNyM2ت U*ȓQ. ־?~@AC[T4剕nfl- mHNq"7Saܸ8Exʤax(X&(i¦Z\Jin*LSI9dC}ʪ.TU],Yئ3;x;vvVMyb`o WKtF,;й̈́y{#9Y u.ͽ ͈ I5@-׹ΣwXJ)3ύCNzz4Bh-y,CL."TV 6EE?hIB&A MZ蚌\`L[_JÈ>&^JVBjhI@uơGiZNHG0%WF[UANS%H}m?GΪ\[c/Y"8zL}"E! +Q)2] -WA'xsL&O3sbZǠS.dƨA{ƠF,4ZD qM)TqHM&("幵*ʪEiqCev[kEDbQ1h>VQz_cnܪA%#I&ňn jUcUP&l>G!J(]PYtI1J\ոV%䨡>fnFbIm$!c1UTNt+ e"]:Fu}`d~veG*0֜w-qO|#x+L.=.vyo!x+<}G2?_M|q7ه|&˿#B/?s7yb?suѿnS|ঽ[RO.? 5N@ NVןgL9yXK -8w?Cz8Cz踘,b +pJ(-P AOc ThUEJ. Zc/'Ú}ͲMҏ/w$jR_ Om ?7?˿G_>T3Tbl]HZ7z !}F'Օ"0Qp_?y|r6 {A;}-Al 7m+P@S:_md 'smSMcȞ\T0Z>wfo gN8I}m*nJݘ%smU|^sD?|XoF'U@ZV9\Ix(+tnT<OOiySXc q,Q*5: 9;L+NaP9Y|뚐Ƿ˰&]]G@g55$š,'7 $PPɄ-ЌnNM-`(}8bk0E ~p?|U3v(~tJ.F΁4QٲH(O` Ec#bo@Mgл_|:;nӀt[o1qOHo0qot}۶nd }z E!5MJΜ\{i.=P3 aM7mxVgcĻ7&RaIR^,;awlom-В[w)vn D T?fA,T f~9k5j:^e|:y5hT<47h-Q`'ȖhYVg_2H8 g 45G4+ F ԦF'B@6@#޾a,dk/9q+j7%wem{+P/򃎬خ8RJ.cdҎ+ E#=H$@,=_ H\RfË/O|)OdDA硢fsݞjMj| WH ^IrbTK/cE5-qCV !xfua.ۇAj%^l>?ՒrTyh,gZf6eX*gLe[+ZYA_Е&S0XP O] !Gxx+%"Z/*W<˷Ji<؅Jy4\J$=GN#a _(".p$@? CjnPv}c\/4/ANE)(7 ݰ'] 1 2\bCAWcerj@>+VdIٻM >q2 y׸vy?Dw+N.lIP(zqtt;'vϵr2y֍TCzxHd&0)@az#q<240 Jx-w>i{#Iog+Sie:ܼh;2L ^)Bs1݂R?=9)=mJ2HIFA2AmsL` 0Ӛk 8bL@#RL vC2ゑvEHgP4 O  ;\($sH?6)Ǖa L56OKֲ,)l,s?Z- lm cF/f,¸TҋbyD;ϊayR0J< 9Gc$z%VsVdq\P.AH.6B"+DTPvK'Jek+ ۿck30e@5ܨCY]WZIrUn`+F*TbbXp/+??C+Ņ󀱫BrbڈZ+m%/Wt1OX-T^ OMD3/U5JNW||S)hWtoO6dS]O2NySyuVS š' 38_!ز n2K8 0bMUk"~wcE{JiޔlONт-- M=+xm6ͪ ZәO)''N4Ibbxj҄{_ic_䟿nrhi飂brpi>uS28 %TY+BzH Wwl-F#Sc\obGIE9UR!}2PŌ 95iAD$)3"tW5ita 1jmJd6^:IbkÕ=nIiIɢ^=0` 'mA 6ژSN9M`}& IN I)T5N bco)ߟ%9+$ROz,dq!Ι` ; u~0|#EHfL+j!hdͽ{g`ؚV䁟벴un5h}8%tt4FV@I;Jϱ5MA'pa}>lj]8!L)HX+l7" K/r"&Xf["O*8PN<7> ?ﶜc+-<Կ x6x[ X3<=G>}Wwmݏo^ݾ>~}ӞP2>xGBHHi/H 7(" "zpBEKn#a$+*E4{5˹0m3}YO{qB‹wg~{qͫwg728m/n7ݛg<9M~;9}~ʪt7WOn=og_?_]t{),1LN7;Cs>s{ۉmxyێ<=fߓjՉm/~~Oct\_OC+w.뱗0fw4r_h ̞Ds0HkbbO /$wTAo]X79;[kvBэwzɛ~`FLiW~=}&6(U݁uQw٣ogL_BOw/?Bkϰݙ m7r`֍{g/LGSYxd~~^qn7{r}ثaO&]8~ 7ɠLz- gR&NKc{|T˓.w;}BfɞOwb\4dbHŐV;e-S>ǝob.5BV8+/aZ")ak[`GqpVQŸ>%ܲMs!>*?2cFĿahi_~i1ÇJKG o*J>$eX˄gDȊ@]9gMX¬]XfuMLvdZuxm*2[v z/bW yto$32$.Ur%jK,= _,ӿKET1]!q/_z],j _g2AB*)՚nPfsܥzv?*>#ʲ۔- PӅon^m:X!DA]d6Z1<&;>UohqN"nN"Bp>~sNt*Kg|wM Ok<:u֗.s$Ex&e֥ifOXP)PS.&-0Yku4tH BI\#"h# E(QH2s}M = ] 6҄=bgX m`?53Y8}?sb164k7a%i7nan/ EjǗ)w9@xܯd[W&~wm2ʪw$|3/gp= U(o}~dyA$u4W¯֍\G* n=g-tm躏{[Yp>Ykqq,EIo?EPsk=nA?q)|(^V)F29KxtMls> gfYF/ \o><|W|I_GQG[ɤD펢U}ssʤ7W} oz4]Ã֪X?NCdhh N񏩋ͺNPkG>G>BWgsgDϐc Sq\ai#Y 24S\UԪ}8mJbf{%tXNh4XSgdi!(ňP8.bu2ϝ:G/2Q-6hCͅ|MT֐P\.4S@q}m@^&CX}O\๠^~vQVAn^Dl5#͋ "&HM&>Nz xY5?jVx]k8bi3O\41?Y+u5nQM-+eTyu!w߉NX#F+*{xnZm֎t1h\m`#M鮢HywdzJ q(mמb{9wb3p5U%(QʕRMs\.Vf \p[`)ެU,.GbɸdNj&?o2In/SzpDJkwxTlN]+ D/~EVLԆDl)-Wx(~rF%Jc9[]ү61d[@1om ofej C7uhc$[q{™IY_|^HL5^܂/"b>N2?g1}]ouh<{8w'A/$Htw%:,4Eu6ޛy{W\Qˍ]Ϻ&eFL=3"1bHMgEQW0*/GBc{9dnǒ&~r7s$npMiojoJ.$R*]sekdԃ Aa1 unT3n_'ZV)™Szt2%ehH:imrkZ+Y_ :µ`ُX:[Vg_?;΋ShdR/W/϶u $3$6&ZH U]0WdS/ όQ 3n=8!w]sF&e$=CRB16}!ܬyeϥɭ$^!M}&XCÑ>a\2ϣ%s pP(23LsҰo2o2Ҕv$Vq-gS]iPrAc ,0ZhIΗCMA<N>w h6ŔTɁōA~7ord;AZAʲg pmc mh;kiP'6lEH,$5 :QzH,QC͜9gs/]11pLl|fHNj%Ul3RV%`7F$c, A ,՚„DPOOb]02XsbwndWadloR]6ӡ/tѰ P;Ő)Y$7 *p!{b! dtag帀[#ȕσ@X+1!hp`]KY*lPT1 :Ag>sQy\;i;LF7w0Enqz[[AYgA4RWb̑O)-Oqv"ȇ3K?3%(3갌v `p_=)ĮX'oS^D? n2@ΰp# ]V}V!9}H˂E,aY2rV%k+:E%M_6S"`%Vq^:ar Vf[Q"q\9,B Irvfa]6Kf< Shx)Vޞ^\(`mkFt# J#3ܸ/](M<|HIv; H3BeZ$5>ExruſG'G:ݽ2pAP?kq0u\-Drlep y,uesOX6$QBģ\BC=NI0TFu 8FcAb\=eL?NO> cBJrQ~aۍ2'f? Qz<2G?I|е]_ԭi!z&LH"\ q}mk7ΫEiLjL[@4ZShfB{s#U~ֿWrz j6;l?[k]t;oPݹ\.<L4OQm&h]^{\TޜCL&k h=maџ,FfA(=3(3m-ƴU"DH$KIx}4OD _|{,HKQCiޡY퓧+,c D/yPZxlys9ԣy uQ 1]ZKwۭ?7UŔ^h~վcw|+Q0Z(jtfl:JJ~{:T?Wn~xO6WΣj}/Ơu=|hy¸vB3AD»~5/;p**7_*b'5DCG4pO7W} c1vFOQtI\ d ;N;ףA(p墊 جuzɽ`+X+3j8:L'r -}d9G=ܦԦɴ5kyk+;dfOKWe00Aw4Z8U^ tmlwSǣ)ӞǣP/_;Hsjta0fNo>$~kŝf]oܫwkV￰fZvgMkmocSmAڌzm6곣>_H/hS__TAJR'BHۣm7hsv=z}]]?kֻ_.Zw筛7)b,w?4m;\_{s\7)o~!Y]~w[mk'XM/aB|yoO~ q5cikZ8hGU=.X]ADY$oy \P.>hv,~6L;Kj9W! <-NE--bے+|7yª ʋcD($ryfsuzS?dUT'OGoΏfUHhW[QѹCDH AIԡ -$Vs\a"~d%rWfc0kj@` xݱsK)ڔ1 2筋Vts rb$Yä׭NNh>OOL҄l-);xH"P*]]HYC}e8mFj65c\=ae0OB-ݽ? Yr Pt^v>tۅEqelT$OEveJ{L>ftD5F@ z Jikcg^v1HNl])L{rFdGx|yqᾣG#[L2SHyŦ^QpBI+`6Q/w-`lqS@eghc,X%<$J[^Sq!6ξtF; IHf܈^TFd#NVif<ٷfLT NXqC1'ׇd=r lWJA^6 \\:: CD(gz̵a[ PS9a)2oFxG!z#F*" [9 FR`xs9&'a^܌*ՍGsЗL|免g' ԾKCnԮWʭ\n:c9:?B'g<*khXx,(*v=DL9 nT8K.i]7^Oɓղl_kdYWܨIx0i -s*`Ļ&y.̓%0"Á\WˮC v\99g»j9%Ey*&P2lfl2!v0iLb\>ýƹb38IO^\>OZPNupǝ,tvBd^iֆ58s˦51}fpvdPLrKiPpDmh%U(Iv\2z-`M!߸$iM |ExB;Ԟ" |U /5.0U^{PPN WXI/;Dy~筋mrlhz<<`0i{1PHEI[o/؉ ) @q~3y "Q()"ȅQ\AL! α=$d>ӵo{DL @f#\ƥģ7YeP$J&7kF6 :iy'M2.iV%MJSyȹfU T^PsvS<[u'k協XbϱB5#t9U~\6%MC?O˭5aM4K۵B9cFګW+MA_a~S?M+/wwVl/Ξgs]_t!$=]\śLjˠpa(޽(ҿƂx篟VZ ~9gi D// '>LL:Ά`!N F1z=]}~3+Z|5zHo~~g˱~R%,R]wt|519_ao20 s2y1O˟ > "7g߿lF;z70m(c]ٽpp&eg?=Νy| =Ʌ9˷d$?!ϖ>"&x>̨xSa""_4zlK):t‚`{|Na W Zp݀E(EQD ?d sf'9kx?<(||3}\CCn'*f:=dS:lsn4q j5&aLpN)nS!WTZNJMAPqVfgXь x))vD[Rl s҅΍VQ#1M9%\^㸲AS#)}0ƞY@FCC6㚷S}8lG:1ǰ R2/=fk 7QCjQRwJE#_~,'l H -g܈7n#܀DI"S3MfDߐ>qMig $01>Ы>jh*{E&fZXwfG1h> ,4%o5mBMK&, oBv!L! &y|b*CW^lG}Rv8gs/g1NX07V"9o=pWƈ" r{fkaX,E^}ɜbŲ˩Ͽ{?~4""A4Y?P< i6O~CN+$ӫ_.MOR_2xˣa8NJ.0`W&0d0 ^@ '̤(:,x^7;][թB)QyuM4R\$~X DWOTlhGy 5I_MR@̶' 8u3RyƼg̙ZI`JL{8Dʧ:(S R'G "X#x=_xuZ\A(GnrΓmt:]eWӸ>;ё,`ї+ǡkQug[LY\_yXmmڐ(c*cBج \$su1K ӑS7hNP]Lj:ŜOOOOsv;PMT8L#26AybgI X%R E.ՌcبT!d/ID:@8Hb ]j \*D aΪ!*GKAS< h~g-c,.b"bZ"9s*a\W 3N1X< 0#80\[$ {Ĭ -㌢=Rshx#>^_W["khUQo NQ CȃSMX9pʃ֜_(*l FTA(mfD\ ZjmFI &)4rfZ'uHbB0B"RS ie`-[ 'P< HHʤ`Y!&h 8HAl@#A_FzNl 4P)98 "6iuv;Z 1T[8IN4n/ꊕxBik/o5E\zC(UR41Zh3LVv9^dHMa.\M#bDaQLId) 1ilpso `in ́.aEM=b-8a72L8L>RSa>Ƣ4z^[''oQ4EWR)8bUʞ@T(M= zUa\$ y`*$.DF@Xs[[)qdQs.qy/YiB^fh4J׏Kuf\jeUM= \CIBU:wkOm[!%ZxD-bCrerV>X .!;uPLXk̿$;$JvX+fj! ~&6OlZ.65 Y0$& fy$?۶'ͮOF}0Նv]1EU'gs{o+㡼f&A^8;tUeA[\GZ:N=kثqKDQcCy͍,$-U7󚙭ăBlfW{S;S.?}ϵkDz^j^q |͊u∤ZȇEuͺͿC -9j%ķ}rٟCdׯP_ -x̐*f&\*x1T¥n5Z;? Zot1Ve{ywZN`w\`wuD.NfՈ&Z|+ewqyD1{LZWۋ᯵; ;ӏjn\+;7XQ1O"6\!Xy\Qٿf(=|r\} *(a &c=SsAo^Jf,&26%# LnlY9nL˪g.ⷐXzb;ϊۼ߲jl^ N!uEIy\^ݿ\8WEfa,pB+s1^jtUA'r@d07tt~lBZoO]рġq9QwRMڠaϞ@YǷUsN<`m4)-{9fzC>7XgiƇ}]>e&#x C+YI˭Ma$VDNYJ$ϟҍ R3[nA&RVcA"S0K@m(TvNl)1Aҹ,LE4<]!JN{NfC|ޡJzS>$f'LI`l ( PtcЕE*Ek3)Gp"c]3[q!hGVʆ˩#YB/}΃ '1"=vx)~gO.??^Jԉb@ k*". ⇋_xc syzZ(P=;$"4 I915ZПR!'⯃S#.΅`t |\5R0瀖0;# ,ĪYF3̍ 2h%<Dž*RT*=a4&hС˥08^=SHfku:_WZ{3΍,:76"TPq QAx{]Y WNQpQ,f%1Gߍ抩s ]G15K cV7.P)BcocMRyh~f.]j ?+v$ ߿#>~@oaݯo 8h(B#ۯo.z?L~0~C.)K^#M/q$Eo`&ͮRUav ]jK/z|a4 M}7#` '}_=#*e`_0O!ur!dK, 9 ·ԌKj[0k* DR7Ѭdad1purpu{l4wA}|Oߌ|BC ge'ʮ15 6 "Sd 9 .΅FU2R$ 6RLh]Ij3zK޾ݢmӘ:SKLKb`֖hd <,i  ,&Tg\ZЙr* (RGrD: e ϼ $ݬ@72N#v Bk -)4E$T`!snNg<A-=·!zC=<1Ψd3Kv4VqT# #j 9J%mahؐ=bQ-6j >mHu0JƊA)iץeJ KybE gkO= ި ?=LF ̨/*48҅HATHJS$J1 e%` G G96+5IgoX5!˺1jcS!;!ĝ?x%H@s)xIC!r-rCҭ-^s qspf5`$@J{!^ B#9!cףW4s".µ1HbKP>Uv0zBI$i*iDDR-FDXVEDUp ;XUd pװ԰j&J&^.7o@n_Lv˜^D|)ҔR7n!8JdB/ܘ Kwy4da#X3) G Ji2MetDR3%Alx d&5QGM%' HŝqLoRɵ'[O}0 `E܏Cot?^AL4}8:{&iyu`z=7VPঋjIfKŽ sZrϗs lzqv;׃O .-yh9 l+%FR )HRd6J.*G~ 0'8sq 4#lƣu?[+ j0%Vw)p9>9h 7so-o T~tܸ_h S!ӎP.a"Z)d51oDĭ܇_-^`E w҆sGق5Ҷgf?5xڧ%Vvfp'~6+ZD-9[wjAPNn.z+`+Fe359,^ g϶ Gz$BDjc"FΤZCG;K!Zy7wRDgr𪸷~p>jcR=0V7Pjqw<@Awxd`,YS(a}6%@Xaw\'sE'spMؑW~zj0_0|^[un3(F#\z?.!#^Pd%hj PDpĬ}F6A1̶3b\9e=ؤ9 7(5' *ʱuMjA h] HY> AxˍtDOL䆬Fpu Hp$qj|3{I);pͥO,=F&Vqtא;T~"VA ` ˄v:fw<&R\֓玢hړEG{A<7Pq o R"$I05jEӮhN[R&]!5Q?R6+J/<ڙi9A}2Rei9=˹7eWd1 w麵P}[b;!yM!h_3-?C%[qRN GQ_;g x,D-TYP盛:E);A3]Hΐnq!OPI4*Dߎ C?Oxw1NJs4ekz{?m3{x{ֈw͞6p+}\^Eƹ*"v1rF\d ؆븕O}V TZH_?}YMp"4m= |֢TaENC96SOi~'l:;c fHEZ`1SP#侾C:%2>>̳Nh8U l#doCJyo7ѿEuSpn v{~TbSL0{:Lrvɾs L:܏tT 8`\}v{^s8s*9˲ °M}OFpyjɸ6r)0SG2BDmDF "ܕbEG|ep4','kE5ghrZRT )g5TJ LC`0(6DIS g9mI3m|(GHDb n"͔eHs5\%OO:RhҥC|c=‰dJPC#bNz4~a^mB:ww{Ѣ `铕e9U͋TrNK/q=|  o!O&aVJh)5Y0J°lȕ(%2gVpfɘٴOiRx:OϞt!RǷ@P‡>K{k/њ xZ,I~h^2+KXT(q n~Y& 㨅5RS5hCreNHX2浄8ݥ +qgU`2 y憯^HS0GHw{; &3YCPheA %scM6k;tҜwT\_A]|Œʊxܨwʼn|d-PYyOw 3 8g$1-ʁ֙e{uyiC6*Ϗ ũuL~ˁhGx-ڟ]LШ/niWWRsxD%vfuKx7 e䉻)?K7?è_3| KlꌢP85T9i-MO퀜+uǾ!O&cdN.A,y)|La Š?&Y,Ssͭ(|IndS!kuv`>vu[kiW;ߑUrIE/cOCшhiDn Me{9MGZB1і&)')dj^5Q#V;i?t%~X5sg?iW>RW ;r. }pAp(*7lA3@-sɐ}j@1#cmNH=:04bhSRY),E ݷ8x%,-RPʠ(" t)`w{#C,myc'62_CP}L/[sJL@3N `80[[E!W{Fr ~Y@Ɂ%W/8?$E obg]+q8S}UU]8 XZZw[.;<ٵ1| =Pj盞>;;u!56|~}ga\Mw["\> ݵNI[.½3MU^+v %ᛟ~lvg[XO0=cp36/s3dλoz V&ƛޏOk6rn'rFPlN|40K쐱2C? .h&wU28Q2AEngx:{Pz1}t͟zxpU4i4,6|/`L/}xx_7?>' G.{8Jmozo~G[h6̺ÇMO\Vђh/ɋE # xy9XglTqJڞQj [>wI82i2,xh^h.$d?Z"?WЗ86w "/f;>v[@)W5ތŏx">I??'JܦD۔(q%nˉ;QX0:(Ԃr.a\`2.<{8ezPFlj<{ 賟/R|5V0fysjxo\x)bN_WkA+Ύdg7qX%;4ޞnSmMu庿CZzb0aR>ENc#9tZ P^[|4AOv,yz[͇Ѳڜ*g?6qYyҊ5LjNHaWdT4F1PLQ͒fl]_L?Oj X) 2NB> Rۻ=h&=3? Ҍ{Xc_jqKrЌJ2bѷ2 l1}nbt6/;x\GmJAǧ{iF/M{ŋWW.ʾ= ֑/lI`Fs_WZnCI^е֗Գֳ?r^-|ν3 rV!K`֣ (R`X%WBKeVGOI!.NMoesrjٙާ=xL'R-VҭI?]YJF:kcc4akJ B ,",QNL S`~ E5VHʍYLRlUoUfGB2$[q7<-[*v2Ȏ?n8 [PB15anl!Z /oy4 OU:jm Q6*ռtXRپJ ~u`aZDХq xPX&Ӡ]u 1LP)2 -'ca94ȃ9 ,v::rvjXAπ'#flˆ4W8cZ))#P P2jqRq\_I/0M軩!J/LK]ita2ZAC!QЊBp%GTHWfVJ%5T aeBY 5Ha0 DӚM,xCURۡ7nJ-y#Uu-^ /×sZQ7*Y AN4i29]rT2ʃv2c*J{d9Q b3Ybʥ?}-05g < |N,m8^7\މPֲX t,r6QHAL`9bm q:UT|4H u`!”qцZ4t2|9)d y5+}=oi÷HE-YVU-Q*N(sF0`EWrS4XX-AOi. ;w L2<Rdh܏7v]n/Suj.8auLhxbDvZl2ymwGl5(_xӖge b HR af%cdcIaÈ6sVrN\0)%Gn)s;$+/VZѓZ#\ʵJ{tv}SQI>NKEW, ZieKjç2(_|Uo|x"Uq+F\S|=_HsoƓ?} H4IM `ILzv`I$>! 74H\qeX cۤ=,Ea"z.=R&9"3RU{^ptS*aާkMd7!z Έm6Ms/:TWsnPʻ|b%e)ne΂wq.cZpj|UGw;LX[5`'ECz]YcrT"ֽW?,<ːK%nVk2my;в.muE[ yf J4{bp͹sϮ>8h9w5F%Ҙ`noڃSICK vj\_iv9 Á)$QBExԁwGgb+ۍTX9m]Q"TmTh&* t=&8T10LRvlhBtk`7U[Na¡X| N1Cs: Z$JY=0h3ө˧ًb.u3NN2Q[aF6j0f17& ]0D~FLƨ6%BWBOQ*,'1hm,RM18`kSǞÿ,% C`C9.D$[mWd5%95%zwCHPs$xFs`pWg4xŝ"ZH'qC唈 vASRN0X*]2d&w?H_N,iP^_ud*!dI_N-WԴ +f7ݓbg_s,Ut}ޤ:d*М6uU2Fyu!)p6͙Fk]U0m@:UD\ ,aWb ,\>ujGOS/5ow mYKeu? K*wME g?7PuTTk|8W%ܓ:!t2ɱsZLPqqф芿PBG,E/۝QNڮ2G4M-Z/496m*Le |`rCENƷY\)TbXeTHjz3: F7EVƩ};#I^fgg3J| zy E,#7 IFhPP8,hTg}aj9# H%jf#~_ q{ -oO?iڅq mV}Ln~z!]mO8Fñ@3n=Vn==9^}n)jdg9G9:cӑ@1!9]= ;F4>3x;%fo?n]ԋ8=W- n‡bIzP->]c9UMUM@vztB5f2OHU*`UzQ]^})d e B PxYj.3ƦMO} ;SK!O= ԍ]ur(Zjh)F}{O{Kqd;N)rhn)nl+"vlu_B<{zUpd,N%}{ (Ȗ޴I?~qdjkOv\#c_ʵ #_1RབྷMuRGLnr$Nk5-!QTi7S!u| Yc{nm4=mjZVC̙ڕS Ǭ`]B$I *$iB@B,}}^`@72-,BBw<5̒O5ZϞ {]YjilyɯEVxՖ9zvٷmoK --VE7ČyUCܚ_ML`g 89bÓ}Zɘn;O4ku +m|m|öb c}.AhAq p\6wpV*T/7 ŸS@\up}}{5._ڠڳ!{}URƒXO,`kRoTQy2+FA&=KyqfzQ:=1} 6K Аa uYoݽɱ7w{UNBYgB>z>͌tQ.Ņ6JHo؛ՔtJ5P$G?>F|ǻ!# ‹ ӫ Ǔzjw)rۃTA@qڎ'ߌnF;SҀxB fǻ㧛+OlxDqmI??dYrỜun1p>(n tNhn,puA#SI-PcH*Ei"3|I|9xsF7atsR8;4"PYM>!k N߆ /;q_FQWЎ'"]:bi f-G/.X&KUbNH49F @sfJ'TYo<>H /UL)FM|2yeЪj%JdI-ab v5Y//%HzPP!xdiH1Ab*xx t9q?z2VH>0/eWBkϮ4{@wzіEL2[ΒM m"QyXa9q, #yMY?ɝԑrDOFu !eBF1&I)21$Q =z~ {<>']1yahжjսqAT˔BLn៳*1r(pGsmU4CRnAdrGŊf#Eq RRCPGWdƠ#F4;lQ}[*IPkp6Sw3 v<=AWħ鿳?8'ʵ@ϵ*//>#by.+*rp\hYi;ȢTmuEJR~|C$M%uC0]{mةfh7Y)_)__KVh C{PBz[>5\+U˕rUNdwܦ{.Hk Wm\ 6}hU~)-wr&1.\e/fd4tWPcaV>|9^OsmOD4{|+wΥ{$cڏOe$mec3mG`Md59xV\y8r#RSa B9B9BDʛ~\Qs4^RӱU8hFN |jЁM8hT?5F~%sؿ~xtWeQ77\Կ΍&KR#i;!K! _\&ւn[yS BT2>MP`!_MX`EOHxgh-4YHTB6QOXe%k@Ã5"hhP0urAO$Dt(!]GHaZݡTtJO0FvP N*oY5W:''8:*d y"z pA$q\ H⼭IU o`#zR%p_ߗìq60+ɧN*4+ʈdِ2rAHo5lo_ [1d1]Nw]n[aXo>hIWAn ;5X1lH Uo-ت|D|Dm#6޿|3#0F{L;B ieYaߺ=SL0鯄:Lp ӽ3p7Ju(.-l5흑;h%iY(ot WX+ߞ>YO/C 2Md[`m3 }&!򤒱wn1'I9/f;BrcB͊,3ݴQkɜH1+0W}YgD\6LO?.;AgQs4rՁ,0HX!YBҒizm.mDs8y}X[ ͙ZM–~"p-sҭ6|>2sr$&a+$`Ed(n4jF30C oN^1AH&zx(p*4NQ#tiԑ!8tZE=/ }?d/^XQw S3b8"p00Xp蓑yQ*̻,hn^Y0sOe {xI˗[e;/j."RB-Db(Et2]TF*-Eh߿}ӛy/ܮ`&%XЖ`aN!НQ٤}byg63aH*DdUM ]ZIi"@vwfDs}4Cn8?O)!{wq;/~_`/|}qgS,`{cODwsA!0\=-97T)?-h# [OÅ>%̑A=3iK.<YN>v޷dθȄE/_C W >֋ƽ8X/َTMP\M)G6:^$:]||؝vM|$GB@z@p^Ma6q*/: _.6Afq˵MwkrQJ%d-abXKhs]B&dٝG @4'^>6 h;Llh)񬸖jVN܂-wdW%)fRabSI!F۲*}>jiZǽzbsw_reI~`g/?7jgJXqU`L5DpWR+K1$FPc} _Vuq/@6Zzo榋ٰnb_~eJ@qq"@Ғj P8sWVj؋c'D@#bBiezmedx!OZ{"- yP/nn&Rw+0 JS()%![VOöwI< `#|V25 J%` ,5PshJe7䙡܂p Ĩ((pKCaJ}"[RR nAP ;G@"Lǜ#YzʝT7Y1L֠L[qET%H^B}i(,-6Z)4N#9rʩ$1eUFQB-}Ij8Qި&pΕSW;;W6\f5Εx(>/s$; ?!pΕA9I #gAП*LХf⎖(r,v2C*-UY!!XlC!)ܮ?t də\+J-J'Vr&^G1K--5Vy0IZ\"MU2d%woa4={lL@N@e뉝.iUIX< Fhx*2ǤZs Ůb`C!Í sgŴLC xo'Su; 9x=? :dPyNa1P_pӫaA^gvRpQÍc3G cm|g꽧` I+[KOΦw[|e3M}qvOlyfZ]>fʫ.R×D<nY?ըƣb1vvv7Yxݎ}rNP` ɹ_A`q/6s뼠맄?nzX;e>_[;潡d%7}3ԍ/S dU W=l!^܌*`Fɬ-BIf1 ,',,t$ݫ߱}p'pƉzgӳNlQgh8ߩg롇lppы+)%@`Ֆ2M1gvmѿxI#JOEAiDtN8 "}ZdZ NPbpo'hSR Zh@0*4T3 $Hbke5TNMFb =Vu{JkqWkǐDb珷ׯat?{AߵH#x*:%a1y g Zh0±QXHgbJ!h1$„C FprFtî;C#t.||Љ*vw"Auk -Y"4I~s: C*S߿Lxu4HũRp;A;h6ϙv!>X ْ3؟k3_[!hyǓ%d%mGK7ȩ],nwsn֛dNHA($ҍǃ 8D=`Hs4 wW^$Jڶ:6[4ϕ? 6 tb̨l򰘧CxCҳ643O.-vw߅Ѿ]B+>eO/N!8$? yBFq6&铅yd:PJ B.",|ڀ$Y~m۬jAy8yE66 cwoOb FlkV0&dꀓ}'!#O^" %97aZZ7 J RsEsEK (R$FIlVDqxv?hSu ~.t<ԽbqWqF#t$i!zOh7oe 'NM՗e 3?,ԭgaWE:5/P޽xw?X'ɗ姛6SEl0%Mӽ(4rcJsFLpˏFLީDz_bI6EW·̈GE30=XL'8&ȣhe~?d/3$=N?S9YH8u4N!:Ꮋ w%%hu[wf'9w$<<ުQ5ݗxkhIAi/|MHgCDBwjjg꛿=ߺY,/sC )VHiA%,0t5pm[!#ױ[t'6`@cѮ > =^H'/t=x,xA=&Ŧ3ˌÄ;|8b E'VԹ^JE!n=|!dJ.ԡ؉ْj~O%UB8B`"@=NPj޲'km vn۠뽨C:3 AdI2 Mvay/#+a0 g`6b'% GC ERԈ=Ɉ c#_UWWuףZ#UZ*KkvT+Ae$yã)HxD7\VbeY/%XX0N9(IƘ`Lq)')Γ  0'`h^-.P$W5$k֨Xp S̀dY(j!0S؁6+#1}.5 &M Y-,ւɗooXUhԂmk=KYJi%DOfɒy:dzoJB-:dd5IR@r9]F [}:f|拃(*??|0JZ}}EE'נGon./""A4)FBϟ^_->Ζ<*//Sz34K>;%B?G#x>6)۽0LGF%D\y= -J`:Jk >~ZK:+$ wlTϜIy:΍$Qe%(Vd4.)p8bcʘS`.Ox?;4̜9NQtfiv؋(mÐ?Kl}a꘣زt9" $D2dÑ?,hg T.Ov8],]lOra^vCFѤ S[LFjӓ8:Io?xpHe5r\\!aY{éN2yDe2(oy :R&S,YR8c_8"5ppio?: э7ͷ=z8B{3 m!5jgEr?G&A+4<6KȥTo0y9.ϖ8SX? }Rޤc3qߛz0Fٯzn`m t~O̧ SpWո;纪ƌPUീY#AuH.-jBBjVZ0|(0 In>h. ]Οu"LTX==j/:PiZθ 4 =ۼ}  aYCBNJx8Fx5șAG(?Fh: (09F%iOv2w51cZĴz}HRB{;+"Q\ΒovبRqzwaZ1O,iѴQE>N\wVDVꤔ[Otέǩ V( nzsQwΦM!0{s3:{st NFCٛU8*Jʻ :wTyBKm=߾M ;=jk +Qm.Cd oJh'v&`g*I '8;*,\w,ZH!qq]UX2qCmvWɡHELp$&8WB\ǛFT@ӉG ڟ5Ҩ,kʌqjtHdN2d 3))͐)0#U*.[Tg-(&O0F[jP7^Yt#iNPC=EMq ֌"`a 0޿ ò~t:8 vR'!_x &Ĥ-d58 _S^h Xkb{=dVpXf?t[!WC`FMNYu:lXr=%_[-xdFGc-|2}p5 ;,jtaBafy1E-:wXV{A%Qm*Ws5(e25u`$fzys{SMPR]t}Z]L8juBrbl>],e^.6dry1$\49߻s1*,r*.2`$w2M{p2oI1ɼNIf3efʂRAX_}>XB !yu'z[N1"kªeƯDž \[TQ>5T˫am(>'\1+R%_S|&T 4lGuMٶ8$yWW27*ƱHXc>)4O|fn./dDCϟ^_G5Ζm3<շ{ -)b7428ׇӨ/5,`-r]|4{m.HfJ1`I5; ݭJIt :q֪hꭳ:Pf4*$,:G9,Lc,Xf\;pu |\ƪXP.MlX4>mvw;]*iL)ܩ[ ȆS|γ(_NB}45=9˩P2#=PڏD{;ݸZJH/}ꉆ%$'v_ $0Iۜǒ#F={Q V`lZ`qnH٦L;$Va2ico/] ==aj$$Ɏx yR-5\O)CTx| ǖiMMy+G31ybq2Ni!KMJ:}1 jIyt;c*IK6D\G(J;e鷷z4י} PhXձȒEÑT&Oϋc$oIN"''naQw~Vdw. ?>U70|XTc>٢&zJ@ nϪ{p cȮW@᱈ !&^ gi͆_ciP*$G0XTȕ#QHfa^ mhߜ6.W^cM9j2MXcw|[O/>;aKk q᧚k;ޮ;ʂ f \UAϧחٜƫ}r`HEL f};cVSUQؑ]WxA5CTw}1]wuymó+3)QX.r6^}YRP.לKZlW%Z)dHg\-`F{%Tx`azvLeBwYfoyI{﮷s!-6iFe B$,֓ i ԓOC?:{L NaVB(ErO3v+C}QXF 0RN i,@4߀[Ni9*`MI %g]Cn;q%ońXI~L$\ƨ!WA~3^8 + (7@Xyr$h+^>j% c~ztr ^eQ+7+a1⷏i@ANn@f%՜+$Mڗ |$i\7w_Us]Iw}X$|%1nV/&|x}f0Y}sœJVч礸T('vq$Q+eu7n%*}3R}n7%݀;uo|=6ܜڌ90ӈi?Wţ/t0#jy3{zr{j.k1ӓG.0#ZC:<__|8hwnw˵ߙ'(B8^ !xa]P$ }~وWLzΦːKaxWf5LIrc|k"$X8iaџh G7J2bD8RuJ^pVJsUFX)1rԞ?x>,NNKblw5ckxkDr*R.Q$vR\` 6jC"qAs6y=U=G(g6p\PQOqT"%9CZGJB;S/xk&\( ;;{'^gVà)LLaViG)W)B v\Oj##A,Gy 'm8Q/PSJ^{3$=Y–NaTp4.eIk<`X-+${-qrT%+9 .# fɂM g:q4ú@ԑin1s.fPI|:Pe[<9QT*je@ pɄ )N`#LqIc&'yO̓Bh,;%0610X`DPуߌ aX!'%#XIPw%Q#oNX. 6a (#Hs0ZE^0F0ڶi&YރPT_ORp<8p?( { )ԀK [8cq ԈfBˁ23JD},-';3: R v k59jVn\0 вDw92uCb,s\zXaua 1Ye>ʵݫ$h G;>ܤᐫOjF'`Ռ錆d4?w*Y |'pi4Op0J 4&/`x5:GulEO:LPܻ§-5OB8^JVj< k+^nHɯZ_Gtj~RͯBUH5S@]pχW^HaVNYΈB쬧  :_܆sooi@|KkgUcZO$* NoI\==Ed#+PZMVTD9̜.PwQIhêޓKm CˏSEPg-}4F/;aWmC>/ۂ3vW:`~c" -u },rB3"%X]?"c̸o Of܌4D&b"/h(MJ_Rv2Wv"l}2W#4bWy&qBL#) 1j@yD\m LJ RR V ˎڬX5赂2 6Ĝ0ѵ.6H"a(4Q)8gKx"JRIwDiJ, sO'aiu8%`6)u; ,WX@C8#7+%0^΁LA,DZdK37V4 L&"mղ޺jL2toF<&8 ˬ2%%k[3׺Q% mN S{B^姌niPw]u$8Y}ajPd ?{wzBҀF<ո -kU'[2O y©ޛWh,آB9`%yR[Ti7y<?݆F> !E^|u U\;\uB i]^u1i<P 8K04.3`:1l5,RxrcEsQte><_<"M9:e`/({.H8vOrW_|X^J̊{5}ݫLjҜ^72+ips^LA |友Օ{iXpHJ1eiQ24-ckMI>cR]R^XYo# Vq8w#$k dD߻ Z&_L>^ՀEi\o暄3E!֖m+l}x?۩7Iݬޑ"fV_Kn5Kܺ3@HYVCy=1=RӁl*\c@%ؐMF8V6H 119+U]v jt)zPm.IaHU" # uic.I v83 r9klP/XŶA]ȞP7º't}@y+ @}}.&+A `W Q ߰¶H<{ P5g!X =g`d)E G"gz>rA[jsEA{G" &ԛ|??-~p.Ƽݛŧ7.uqn2\%qq\n?P/*g-NUbҔ 2.'w }ƴɝhҧ=or+El>M֟[sG 4'SfWKj*gMip^#iDZ35_q{WzzpaVw+tb-ԝSi4$j>iz#,AB2Ԡ[L sNvQ6eP=1@4脙:aj$'DN)e/e$N;=]ʾ DiP$'%5[ª A©@%Df Նb<-.su7[<&wK{x2E;f){tY!O@G)Tw5H>QkêAJp`|d.@qZU;3sάrSظ9Lz t@9S$~j4r0#" oicĺ ໕6>Nc^&=ݣ r-J%%Jzx-Jj",~MvhWO.n b,<ofp>lP7pCl+a0ba (dR"[>"A%W0 + sG#>nG38&.ӻϰQMQ6fZN"T0 i/8+K* #(+ĤʡR{3fl1r٣k?8&VZ Y* +޵qc"bg 5yx71q4,%R3o_RRJRUU$fU~yH@Z9Q% x~ Z q$i.SFvɔڻaFM/. #+DI0ZHVG,0l_${?O+dkyG\rPt(QhWߣhx @5OmI%W"QhW|cĐ=6 q9[dT ʵ(dOFZ6qz[ p 4E lZ>hNuQn,F+^O24j#k|т+N!O^<*~^1N4 aBd{1,'2ׂZ3EUBvdBGr@sP2$H$eį뀘:zII?pZE/ף` K s ĈW8..?`S+4(ʿceuw ~8"MvgzSK<7v*%.Jɖ)ԽnI BғW7f37@>~dh_ެ5])HAE@/ҚQ4PW)6(JER)+%[fr~W҈D ڡ 誙Xsx*]v.Iۤ+.X?Tfbd'-v!OQ^FU!}XrFTɽW86ZXv ^i>b8[X$\6y|V^`x/+}ZNwNˋRSU&of\֙_`4zߧxU&r5EK)CD=4o^ryqϫt;JvpLaCSŕ:#5!дOA7,|i*'VH9ijGqt}])Ő)-]HNz$U?bOE;j޵qՃa Kg g[smOzP y>PX*yFPWR**aYkn׳P3ieGߗDg y]aF223+y\j]KPFw#;g1T@Xe!p1WMRݸood=/GCf?N]=XY{Ә c*ɫ0&cr~֟m&a,=29h227$X8Fgɯ? Ay3\ޭ-h~KL>&Q/}x> 7'_a&gA1v:`J|,n&K%x%&β۸ưqyok~0#µtG:Fliѿk0߲E#cƙ/o77ߦ|qcV!ҡzhhVf"0⽥6 _(+1f _]w7 36Uww-vAlQ>RGdحǂ&eȆ\ݧE!"44ߛֽX.F[i0 :O^IA(h y1 CjZڑ}$Sa'JK` =SSĈZζh4}ڿSDX:"Ca:ń_WItpzmt]v k3ҷx+k^d!Yf@>jx猋3 % Id)ʕ9d 㭠Y""VyDeI?خ,fHr-Ҽϔk <.q-yRBw'HDӻ8mSt1VqqqǸ:^`c 7 qBQ֟GWɾr.Ϩt*y^rjYl]+|] = 'wf; $ %VEoa 2*Fu?v1tߕvUP|~ԫ鶖Z0=o׮6x<[,$~{-↓m8 5oﳇ/o/.f\_n?}n좜/N<f~oMIjpyF[)@CF7ň͈T`H^dNswM<Tw{겉 Gz;񟙗DL,*S*H*K 1՜Q ޾M?{sy.$`rC e)D4pBN &oQ)&}o}[RL=ۤXcIKeFQb2)cȫN0A%ikX3ɠL5 |$ 5/N H(NEB%p$vx@pE]CN@VJ@g6jxF 㷰LEFXN+~R}-r.&km40r arpR:cL{݃3wBB={_R~+;1z8?tӺwJh ).Ρ!?Zu11Ԋa n9cTIaI),o B㚅܁0vwùH%;DҞ]='IH4@y u?^GxG.Gr,(S+X]x˦4QkÎ3BXO$ڧeS?r@ &WRqSm8alΔVsjri&AZg.zPȝu~` ɌC[&'R ˀ04DT Ef$WQJcK,^r R "oۻ?C,J#jʿfeuWq`t*\MWzԄx,,A!8Vg9FbE#NHNe@kɵdJN˲%QPK`&(re$㒺 ;, xi%KH=wRkE /c( >^N!T!'.e891P=9x7!? ) UMv7e}8Y{E<2d/s}eCRd72VAث{^P~sl_x54%dPh"IdPHF!XfF}DEao'1~8YJt9(=F|Vxߊ\#v ;*IG$j#Nl\#˅J`:HJv]6#s!almɄ#ػ`\yiӽʟ٧ -n&w:Χ;q/.FēNu(+^2~..~$R } ׄݮO SLYni{H+3翽r5}|\WD+{&DqS{zr!PJQd- SurD)sq>gЗT\ې!t}0!U~}>1*QbG`'x^y0pJ.\jxȢi? h򘜼mo?"sؒMrJ&S,L\iE%d_7b]m$(I=P.jZƯ3'ICHjϘ!E s‹(|:2R'L4=L]m~ǻE1SSڤyI/eV9_*dkSJT('dӵYn:NVd o& b2~-8.-cz8_%EXI'BO_dDES==  òtWUUnMe(uQ +"0]VOYrROˆ>a kTøp2B Mk&qOo56hĔhz=`Dt)ქX9KBR B9JfF;82;wׇV%^=M/(#}NMG}Ls;$&!U(! 3!ǂj3t$evHSQ>TM-tD'㨔yUPпfߘx<@G13AfB489CEv++ܫ~Y݁.Xq12RAABjV>Ҡڃ/ǒRRszL 9QI{xM?v`F;D7y8Qf/>1yFjxi])M> &'kq5%\(&zʐB #3N"X0'V"gl(bʾT-Bԇiљr ynA8l5TbGK;>ZS.ј`d;:RH=0b _]] >*Z])19l 9 HR /J$  WFY0na'>+k>)#ۂ@v"cr,r.l6`4"(q`v|\;PRcd\a g `h 9VC.tl B+=nU\È!f@ GD&-RF}-4תaXaEQ(pJd5kF֖0\VpG69 K'Kt%<;q]cZ%JT!KmSdjŃoE-xv|&0$v\8 el#B־Tp8"g JhyUn" +.ȍBxN|9B$5u&887 u/vƒa&D as*+a˭}lϽDLF=KR233vYoWCKURy$_e˳ثf3a˳ա^ 9)Q]&3fO1$/,n2- j_/va>]RTk1bm݇ʷ\AGf`'~QKq> ϋ6140d +b!8gH m6IA'`Vh,n}m6!6 5C'G~*C$SjCHFV@ ԜjB x׽5=I2xЬ #[(Qp+- .ǂ,Nj/6hn%,ww AmM AAԷĚNd}VJP)*#j4m>v,*6LO_Lm>r:Hg_\ [`F ++)"0()W0sǿjAvmI\~! Tq-4rN5cuzv34- se\ qM{aϻCǏ77o #"]vs퐍:uЇ9]Ŋtu]v7qr+Zͯ?K=Akث n0l7Owדmm?ڎtwβ1i>f9ĹWZn6jBBcRgj*ndG;f!a~١»gxg#ݖZ!L4=lFv/.čZEP>&R).ѡ2F'n(I~ӥ>*BEA(\pyk1NDh [ A[H>7:VfD$tcjayёX ;an:ooULl{R]7h zܢ:zݥh^IE)e~wFSx\;Zn, ̿WK4qKݪu33y?y5V[;ϫ>ANK/&.BUlQ99$חjE/}粆I#C1WP$&H=y\l ^PLOmC=CϤ%ę-WNmgT\a\N co%T.Ϧ0c8~(BbYxf<Ɠ/ԝeD{Ae(dJқjS! <`~XhTShE%ـr `bT0SX\F(3˗şQiI:m^5* 5wz 8yh_?IxM/BA'D7ѥ{7_/l-oF%$1agT#9 5W{ ޟ]PKpk _'%+R fQ'LbӴP8/&~w2&S ewE_jcrjt؋cNyE|/Mg9,zWu)vO\tg^/57x=?ыIgׯK1ܭ<V 5$!o\DȔ,e|EXGڭ&ISsJ`]ڭv)Tրq#S$捬j7h[] RDUrvWJnMHZ2XeV-nu1HcTn7z./8.¯"[0ĖQZgrxj( ۽NBn蒔uYRZJ{]DލBxM? j`_Wyof/?(?_ }"4be ''/ q! |`kk,QF`A$l195yL%b8<"*2-9D 鯼]7ϻ^u5CHiоuFrcKBu+/h.-v0ѲKir%gH"T*|\6 f5D+^@6y,`"gy\ YcTK0]3 Z0|D﷢ Ą Z%r0J\`$ arI6@|NBq8Z"9=\IjY*&(a͔y 0@ 8V*o=Z.@)nHy4HEp$L6W"ύr2%l)[ Pss7d֏Ei \>k7-WMnW5P֐!8iJMkiJ뒵PPgUfbьc l~xi^3[D9AM[|QrYTޒT"HF:F(tc{)NaAhU;n$ AX5klAiٲ$8-<$(:❰'= ,5K /:qhqk.[MkwkB!$hjgQ_E>`S}V1ѴZmqnE+)~䥔H<&r0)Hεю(FDpa,.ln5" >*ty⇾3N4՝]֒-W`>$]q [KJR6VFyՎd1`fR,ՊOS|}gٟh5Qm)(-qxZe?N!`Z#vm*Lh&L$!1$0v&sh[^y8Ns٥CQHI忧q@$H1==_wt3&f?eP$ϩ9q(ً~`Ly@)HjGJ˳c, HqdCd"I],c*觨uUUt8cZϯ%aD)IZR'v::raw #c]GG/yf~ھ)$ݗl] !啭V{OFB$դ@݊m|!XgZ"q>9%"Di#r],!(LwJ ݞϭ3[b+6 sOjK2-rUBu6U,Dʳ_m1ZH v/odS1A*\Q{=&noϢM՚wz6+j'vҧ9Rn b;rMZ CFY(k)2e+BQbHlt_\O%<6<]ԛ\t8OVI|hߤֈ_1ƁC#޷r!\Nr7H Dvk'-J4p>(h'9u&p!Mmn.{N͒*a"g;qňj;1eE]7]I.s:Sy͈:s$NI`k砉d `0,I`12* +VVb<^\ 4 g,7ȭ0Dy^:9@7bݟqd&/U0$sR1hGZY ꑸgn׽p:B/l\}3>DЫ(N߳Y|d*K$I+ZSzwC48WI%fMԲ!D\;ƨ.&tr 0#T*}HHl,?aڤHK 0&@"2fT?2JJ*t5zIJ@:MĹET(b〧|Jvb1trf5TqjWegNgߣ1jd=<doQ#{ppuA^%QJЎ\eyllO~vE]v*7R4c )WZڛcKǔӋMq2iRNF92T0JGAAeLHWn}817.[{"eH >\ L0UF#N"\.%rJV"/:ss6 N uWvL#pʜl/VhvДw!3B\JI:;R-jӞ[yڗגRqp%JCjA gF[NaL1?N\lU*we.f,|%opN[v$z/M.U=ͺnb||rm6"Jh8Y}$d1"CA J-XcaJmS[KV`t\=#QH/=`h+PLtI㨄 za$+fuk$<| ^Հ*-8<_t>e姯7 z&g6cФyO2CcQll9-G 5יd b58cʃ jG6_#+.W9^az>.ooAx(~|;h:f7рe|1`c<<ƃoLKvĤ`ϙL(:Kq8R;҄u$!JZri9Â"vCZH^p{zϖT~W ™Tw ʹ;7ό!t_ }8QʹQ8⡃(v0FhT`T G0uRxFH8Y@J&acX04 h:PpD`AAvن:]uyt2TT(޲[VxjTeUT] 9:G r;%-Qk$HMm0adsAU@T۫T=ȭ1MTb\ 'm(mXOd}s~oM.=\@%!H`6hSu4C(R+ nؼ-#uW+|z?GS Wɱ f\vke (3!Г@x z`(spb{6[D/Mՙsp<ʣ/3+xmy!soR;Q@*T#.&$ԙ8IAܘȅ9 ˽yN yGUGWNDye1m!d^IAa^in%jIf`U.5 Þzư˸U [b0 9WN_` ƽ&N(&1>ir; Nf)/#GQA$P# Z[l 0SłcFXQ&f t9*ہwRȀĀ qރl$J!'=; H3ZVn:{0*m`M1?auSI8>  ] ˟۵go4zzA !ْ3 1\rf+S%D *}xLppVJ" (Y+XsI4g<yL N$0vyuP%fO=TUl1DܑGlxRNE,7PV#eƘƘ>՜oD# WL߸&R"ۢ9ŢL5XH=[P%`Q`Ӊ:׷{&Sn@| Ja*.7i`0BJTsũDTQ19(TLFhx6h N$=[P躊DG` .5p^NB+ĜL 8g#p\b,(ZED =\fjy8&}T9SPs]ZhզR#ħ(Z.}A#@R Y'on]1-3.xR hؗh} >L>˿x᷃嗿_>H= ci1&wGv e-` V<֣:S]c=ee9x.Gu?κuYs51g)9ʭgǚP}ȪC+ݨ~rfDUrڪOvӺ\0ڛQ~Q ?$Lmtu7hk,^+y+LG}=lizF|o\@pR/ȆVd%Vu.%C9+̓`ݿU R̈́Dc]m sTw*y:꩙&ύ l~nEXkTvl:qX{쮑UرvQ+ N2ؒlPT)A5bk5hbV]xe`B%^]kbl%铄P1a.-jaCÒJ";)N|>eE vI'7OWyYqoZ'䮚:5YKQQfQG/OI/"GBѯWFw?$z16G>ֳ:qYbʨD\M'(z3酵yo%!1&$SZ?qxPʟ&s@QL2*LTCv= 7[dXN aGj6!\[e@!x7Nɂ!pV1^3!oI@| i, c1v 5rd5# S#+Q9^aTNUnW.xEb˫Z'K[>hvi D-9!9#`Zv6YJ :X+In} !#4 h/]rM0zwF7) zv&eŊ%hYcK,T:GQ? k4km1,[f9txJ?7`[M K-1S WN%Xg`2 Kν&N(&1>ig ͫ>r,bdhsr\޶bp7͛_[~},kqlg hr׫Oo!fo!fٺz@K\spr"Yр>Dy3؆,Ȝ+3Zoc,S,]cv.HfKl5_m1 VhDg>G=oqQ~Hs6ISo}*~ a)6]pONB\?=wRbDashziqEOHi@%M4p}9#B$᠜)V7,GZZ(M$U\}X!NHqqwљqdyyT!QQt阘2'C N:3NN"s2aεhAJ}_%<5I'艛,(3|=\?Se4~VhfCK`Ъ]j,-JO^wxgA SJ^6s2]Zzh Qb{+ . #z۷#vvLu{}ᓹ4 {5|2_+p)o_h/fuZ5F)Zl~{3{F F"*J戮Khqٖ<]+\xF6s%6V*\g|8 7WNxxW[j޽3YУHj7Y[\+si.Zʂ. Rkܰ\]WY'cðB0ATca7엣dy< OҼNp[Jcmpx%Rh:F;s9c}v__% DRQ{$3:P, ON賖lg5Ļ^w`-n%Ry];74runj'76pu|?ys?W? ,/-nn/ܚ%,? 2 ?]+p?Rظg0r\j)ʔt9U4M nxLww &Zy=kUJ?r]f|*ZSb^X7:GnN;rsL;[yHu!_&久@:_u&С\~G]$kPq&(FyO/Q͔ackFCc/IgCd%=< >#zj$ȿD@V1<2 ҕ6 {z.󗮶I*#j{0ׁ8#jZj[pbTVu$Bb y atwZK(b ~͎z=K8-z~l{sD #Ȧo.2J'Bh: &ް-oؖN[w]M(=C?)&"59tgӉ:Y0t^Ҫ/8Db|))dȉ?V2YcBXCCpSZR*S1KJjr ׻~ź5ꖱ,`YJ6,}Ͱ(S\Vqhth,0}|& k!+TR+$@B7PfT [2ɉ6XW:0unR폢RkɟBW A#+3!Fj6q/D[ 76LDVe}1?z(˵(K2HnI8„#_-.&+ j"*zM;xkJ }^M#P~j)- %aG9w( ~nJadY0F5LX>?bn)`ˈ!(g kB} kzS`(,H*oiDPQ#qiX e4EM. fҍi&c "4Ld9 ֒3B4&qcޘ7*u{!^_?n7&nwvC+|"vȻwvXM|{ۏzre>]3ncɏ$,W쿿8>ŋn~,W6~j|྾Ho|zGpQE))?FqרI/ >V/4h{?ߜ0%o}Oӯ; c(>(4#54X0!4( ;x/),7sl$9j`X&Fc 2#LN` ac '@_ڂ2l-h v8##Nywp(RR0ЄBSgwDûK1C)6˽  . ! ÅYHd(hkV`o}<6t+'a ܘ?LvB1I?RYl1klUMPf1ds* W%l`!rgHDw17NF@%!6,-50[cX2LLX0")-`MAl!xLcmQ@Kg ¸ jn@':P[ { 뭭cvl/.pNrSw ) tNM BZR9#D\[h汈%El24DpoÒynj!h|jڛ~cO۰mؓcoT3#t^uXz^"D7*Kq ΃30mCNnGE B #:tNz:83Ie~=?wJpUeN)ɚݣw;=B |=2Via%j 92a j0:FacD= B;GJT(He@#T㱫=OM{&9k'sIDR,GlP E\pJ !a-*vs1 S fHZsfx)O${1o!h ؂5@WKƔ"&XYԆ}$IfĤ-k%GBB=E‡L(k NYYMT^Z&*/|u7j3eITr6Q@e-瓋U(%fhN:fH2NżU!|Jٍc!B*͌=Nj.pFD#lqGEA˓3Q3S3! al tL6uYwH48KrNjE ˘H&,Bu>d_/n,f,]ฐ3={)g)9ci@Z 9V&(7xy[aq*XXCDy<FX #DXlHJFyf =w1SsΕՔ{C-B}NX)a24ʭ8~sq,Hh%8BzV`fFX)F:T`^WYĹ#y|8j?>\|5` o|@.ՒbtúonBFIv\|ݎ+{"9pUj`|s\  Zfr 2pHƝА-7ᒒEdL.NWSlϊ};j~Y#] F5R?N;.I'gdI稢ZP17KPPhN4)’i)5 NxNhE5`aVPS7iґFjv)ZOuOT+Y5)+d8Δ Xj069Q c/+/s&AQ<;e5#Gt! zU.MG,[v͎K!Bw]>?!RM72*,G)a w\ucKD Yhe5].O%%.KF3ƴv/r ̪ L:MHYMI6ʩXPe ˛hR!a9k'Ax Qq$j EЧ 3k%Aރm)mJ:?C;0)2j"J&sxTo1Ըј!1{OҝH CNfyKq?J!ϲ;KˡbgA zrF*$RcB zfDGF(DH8{YkQAY):Ԝ)Tq"xJ~qst@Nd )C_Sd&)TԘz`P7^e=ffde>3X( zU٠X(^y<B1sP fD߁DlA+IBJ L"!Raj-%J YUAOq-aa#c F-T{iT!JQeYy**NI,5dS_gہpZ:a!="b[14`Ъ@R 2FZ;[a%|P{V Y.)62(<Fh)Qa[!Xm(ImpQFV'^@%:!r7DTTT(X+J -"18ʢF֟8ejn Ʉ~Z/ S͔ldRPbgӭ"x)"KIĴE(8:Xb;xN1j89A;|Gy)N-['b7I.] z@p ҭaoRHs&;i͜'7xHQp)H88Pn,dBzeLyf 4^dP9HX_p H0(&*0(ZB:j`p""^HV^9:?K8OD4KV`Ql xO5a86U'ΧMy(>^4bˍ(:"nmr[p||^ѯ+-]ڰP-nn,uW|a3_|L1 wqBݝ?,qOq#4:^(vu^U^o4Uy;?P묭?i ^|+WwV;_YkQyZtc (r1HY:ctn"OɆnEҭ yK4kM5k5OZؖn)H\ RN],F![~(Et+BZS$ͩ[M4[Sebt2$ݎ#ͬ[vSt+BYSMoljOTҳ)vST7Cjٞ/Sgl QdRl'턂;ARdDuk77_g&IdEj%>j"8a)w K#MX>RH]5<;7KӧM4¬dBjDufoa@GgYGaN v,v= b`LwҷYvr9 (= ajTgc u-o+>USޣ:BmF&NaA@߃0$px.FH{1ahcT#t`tR9@H2d J4%c1w<(i搻$VQh"[mI1qwϤHy ]Ća<ʲ$iGɌZi,WnOO7~ ;>0,hW1o U|"F^dz\=g}̘W7 # Z| \͢/Epo$WL~7+ɘ:k.Ս719j{xyyYTW롯L*~|{s,ql`J9yS>jDTX9'*3p4WZqf GLˊ)dۍ:>)T1Vy,!@R 1% -ל:ki -᥅d ;}gR(FNرNhb|2W3x`˜1i /DqiZq[hylEIŕEI-|MHr05 ~C  q̈m!тSvEe&Dh3jlBU"j@7(L+^p*!3rr3X q־]WJԐ*H:*H깪ey@(jP~ kݰ'u'S`XK8:SH"6aPkhPF"RIÜD,h OJ@Ѩ^҆hehKvBKp"%7"`݈TυDj:6"U4&)Ĝ he~g=,mA>H1j#׋-nÃ2O˻}|#fla!Af9v53w{m"cr'Y?}K+WO4ۇeъA2T: oţxnexMEZ^?҆قxkt}`ϘVf s8YcCh6RۼEQV#.yԏGpL`5}7xb"$e$V۾]l*``2XuH>nVQ= `Q%\t!&E$xkDpÜZ3,@BeE Yb<3vo5I1U߆ {WuUfgZX ꃗZ7A5mut4uu_\ʺ0[i*9>ӰiEqU'(±RR(,e( NS|hʒl8 HcyH+#ryJiG?>ER0׭n8 .Mn)\ -!s 7כ~m7P痢Xݽ+ux_ݤcoo,n_iA}pgnfv''z8rPMh^2_¬aЯ}~kH.E++nfV[2c"^4".8P tUPFB$}Ŭ0dco*쩋f `*s@``Y~JIx ?jB2~,L :X -j6|q_㫌xܐٳۘOQїՕBZRz jv!H`i5Lyrzg|r=rwVa< =S!ޝ=Et?t9 cLr1ٓAL'WO4mO&=jO(eOv(ँ8SdbY|V.-%*Ҕ,Abv+~WJiW6bɦ6=b#-F$1ZEv Mg&nQuA!  5S2uWRxFH8Y@J&acX04 N)ҳq3jq 9\c]! r=ZJ3k}"OҢ';^4R&eMѯ 4QitVylic1֬`ãLr'0jWc%W۠tESu[mߔlpPiï,%dh"&~}U?JyCk“*r&b3ikH /w~?f%$\$nf7~՜!GiFÐ=ŮUuuUw=@ Y>ohc oj9/oW \XJ/*6& P&޾&t33DeַMa{\]wœ=)cŕx#:z>ȸB).{7+3w/Ԍ1OsD9|l^ ihz8f tM{pA/񂱮:ye? :+rr*aF/4 rbAT3d>?o bzAϥ"*P't/T?[G֘P3 {-5B}[ԲvXN^Wr\\4% -.oIj76,];v"wD,r74`T )<Ib$eղ44RS˄$f1Q{{~G*,<<ϖgj8v6qҳv[cZ~d|5diê8U/pbqN6/ORS )՝*5-"o)FeO|!mۜtNڄ.VwO6]ؠ6u+1A/ @vvv0 V&l0kE3X8hVeBHaUL(mA±SP"J0~m`(VÐ$sJX;^OE~%\z= Px(- p\iA^=Ps:ӝ NwOw  Fs0f;cK)Ŝ2 bӧ;-kytJs@EUn JXS-wXfBT9CD<,y#pcF]@63+Vv/Hph!@)M5HE2 Iu.-Hs0RM54eRQݚRŏ B*E,$H N0CE9V&w;]ec- <-l\P$,^ =uL}&صQ~Z@I{X|Gƥ/u}~tN#f*6.^+G4JewHBDK1 S[jK<^ߙ{s,M0( s,+ yR>!`Ri&Z_,zOA|rsWz.L*sI.As"L |:R嬂"HmjY[kLJI Kg- 2*6]9T_.0|?<. ?ki-ugg.?z ֲz7 P)TNrJ i-fKLYy%OO~yS+ẃ@oo x\&eۗ@kUh7E"*1=z^W߿yQmJE+v Mxo 5(vO5_ R{k}[cƖ%2f6X:ØS`2^WO;_ylz &Gny2r>/lήH3#[`_*r<ǃJm_)Ƌ5cM[ {zTPa (rq>th1Zi|+*ďq&D.q*[*';fD+OF.BwP/?ݓ9 {a =;|24de=f/M,*uQMA8Qܜ>h?=T;)_@UaIZFdIj7k%儶 2p :'') ն*8ged-'lhM-}:A[ N*D.5.k!>wZ= ӻfvq;I.VC_#§&( V7fRL-+;iʶ;m-YghώZ63h{<#Xua?#Dΰ#Δt8#ZPk{lXu>_w>zv}}O63h}Ҍ[a<Բ޹znfPi;b7l A\E8WScIOte=pMFŭȿ7b"b:wήpHvnCagԲXM>vnPijTkByuCB+BTiW_o<:CC+,$,$Hay1jC `?=h𾩃tt l=xhwpQR,EajP=Uj@nKRO9B\۰ÒMhtӿ5!U ^րJ lCׄkD/AzjkM( )$+%o=Hь\4fJ JVf2q/qXkFPX+y63<҄QWQ?nx˥[% rmǿ1}~8$Gdxp,G] {\}j硞?orxĥ>p~Ll'+FVB FA*.( &o Ȉ8.ǡ R?HЫuj{zC{m 0pk$tYL&XJVc$HJRc)^V]ױTzz/&BA "R:ЦR1Kq;2$ufM%履MrRFKDRkdDSL 'S(nTWb{Sƌmܛ[(z88+jY"/??`/QTL{@cע'1|^]skwv`74TSCIry jʴ`T".C"#TN(~vXS]lL'&IXX(t,C Dl8sEm1Zj(ƌli'IZ}0_UЯ 'wA|PfJwG@tY?(f@´[:)Bc+@\Kmn}ڱNlkā,M9O~wgl{g3wZY;s{w6{Ho|$ѻŷN3D9>ftu0=E%1]>?R9]YsG+ N4TwU*Bfccg׎xfBQDiߪnhMT4DY_fgeAgTM7wuH߮ 5h8$U/u;O:EӿvrW$t,1:*c!/DؔQ Mu$mWNݺbc:]ƻ!ꔉѼ[z>,䅛MLb3D#2>-PBR/P@ +1qw&q׽u۸v <*;`7Ч ARse~bhS53Yq%41V¸<\1=1 .Y6EziL!# Z1c 2ǪRL\SPH]/o@TJCZ;ȸ.5#x^7i5<4b,"M'T!E%'_(f*Ys!EmFA#$2XC=bB3%ڨ`̘h^1jo2ިCj8Ybq Au^Ŷv #o6Q]v+)iwx?}yܡ/0c^+nK!7!DsKa`1ZP86LvFNK)kZDGH gP*M9}UPݰ>rFso RD`KƄ8)(`!1Q.XU8-%݇B<_>,#x3)jVW_j,zr?ߑRO߿'}v7oW-]Q Z??ˋpuX`je-ε^,*2}I}4+꒭CeqpY|G#LIE3Z(re#\wO^_/?}o-__㔏>}]RZQWg It^S#=x+Ibs㗩Bx\iQ#ASc $T)GXu\a)a1S}VAVҥ?(ƧKn?^(025犐Z L#pCS62p!,8m$9d-ܒ\,(n=Dg7!X?xP9i\ig>橦)`p:0Yء?_]Qt9ۜ(*0<:DuŦ⦑I$Ӿ9#(:A HD&bb'@'" 8i5P,gؼN`aY nBnы_B1Nj[JO.ҰBvpR)%C.Sq]NiWeapẠ~^*7X޾AamŽ?ۉ(;v]hg?Zoճnxt!4"ky< 7W$Y6ܬf D Ng*.][NyO<OLyzsO$䛘l5շOP0E]~/~@O2Azj;eVp{:!Р?Hxx# dW#p? y2 9) VV d kQ:J C"pɒ5 |JRި(„x+F Ab%Ŋ{Y{W $(eǮ 9$XF ?Ǖ2IaRΎHp7xl1kg>JՑ-o[HH&aR\i|:o/k[BYm47}o[`GIǠGTJP8;nh^K!OY~b>V1N=ֽz_zYub|ۇC?˻7o0jDꭒI ѥ+ƍ/q.^߿/,- -K:SU_ _\Lj*:Ўb)Gж)j1K62Ʈ}I^LzH \1NIbNr"|`ζ1D*6cZņ.n ?>p5Vr7_rhA޹;qv p| SL)2]VgTWKXmSt]2's?@5Ek}b.cx.qhRtLF(](H%'' MXHYU׷{HHhB"/r.Fr+qo@ bSB{} Ug Nsjs]?<`KKlV[sktZi]A5igZ[(cHR.'\ݚ*ŢTl3f "TCQGZ8@h6ub>1/ď?40,h VafP,"D7BZ@9Ht%d $rZ|KMKүt7hT-ꆵ3rV\]!'+V$;LZbvܴ[BJeBo?Z-^*H*/͹G-^wn-~.bH)%ZdB'bzr/|+Ȏ)WWZy8VŹ1Jb-4*@m߄6T{@XMK[=da2UTa"OU0 D"z<( 5Ί[oB Ak.2h9X+ >G7Y;,ƒ{'$?/4W`J E0LZPXE(`)h$!8&p;U#jUY]5OyBhȨasmXFOBt鏥Ou񓍈u-:`,ZOnͺ#!z8~WShQO(M5T1{T[oa'SeWI?뺚_-;)ך:Mk(:=/G:WjFLMgs:.IE{d"\{PqM:]D^uкޚt^[%|m$8b!d-f*9hzj,V{?g}a&\ m7Ne0 z}2D>׌gz:q[&{Փ;~ PT= MV.>w7`>c[WVvnSۑv 41D fT,)]l^Q6~/H:,xXHu6}yd=#u?sY!9K4 / - )x@bdlY[{WѺ$ßda*R8˜Fw3RB $$2IMq"1vON6U>bJ(g^xskG5xZ6U+E=^f)dM9׋lJڢ˓Ԗ,yT\<,[~O`ǟB,1e9A*C̴=[)b+p׷zfsR9TZʎrU>̱*˺||scOC<\3F=qk@xe;_n[h#%)rB6*$RJ>-U9=n-gwAqhĝmPVZ BE2̵ppA#*ɨ:35U\!q^Y]qd0wΨa [$1b(E Z.84{`:sI噸#ZzJ"5gvy.)M GuvOM=D6xom=̓$4NǪ-n;|AxͮƂ ΰhjz =7%\֦ F! y x0Ԏ81cO cjwQӃ'~0]}\wK;RY -bm-NJ:4ECLr>,䅛hL'z7YK>wAt|F+vyV*:w+ ׻`!/D)4Z6Ѹ'IONz-ŨW>t@w#SiFGA4:uM2d:V19Pl*>BccTqtI?sWV%IfV N˪.m.t"C/< U@*M: on^-}6f͵f8Bg@ϑjAP1̲TqgJgO,4%儫:j)?aEEj"'|%q-K)\v5 _rJ͝CVv6JW>]덍:]Pf撕IEhoxQCn+)WqT#W(-acĖ=HC> ;^G4B\/j#Wa.(=6r]c4k#QFnNƮu(7o쩵ɖfmȡu`bVmɴ-MX:#ֺi-w-4읅bjTc#;(3Z7؈;_shҐr>=>bBaϝ:EhsM ҺgXʰ(t`h)4SVm=%#lN>w($r@-ށxL=sE2!&>$hk$ze PHڮJs^5]ql6p|6:1%A "&mb6:­p^QJR*㍌:j$2 rFJFPFhm ՄD׭BF SJ#v1 d'`^`}z{7}ԙ$w.I=C%;O:H#'B8'`Qn2Ep&DI6v w rĝqvZK_Sb[HGF w?mO"w`7QGlaB)zFnX;Z1P$ϥ7J'xu7v kI 3 3aio83z.t]0űL'WPL.kԨg:4{5P':Aa[-NZuSr1,Ou.*@QΏO#=V<&T*4Bc:ӀD]4y[%m.:mᚴ"7㨙tr<-s q.^B߮BLa\k to}fXSLJ=MD_G^PO!UC, c;58v^=9uW@VױqKS&\zVX_Ֆ&^\a/Xj -L 1I萧DQ-Ih\)IqəViM#RYpqF/d)EB*hgrhѸEl +\t.~fT}&$l7޾iGn#wE3,6%󜢧 ޛIp85xf2V]J금$PmbTD[mV *:awLsoRLJQT() iɤ:*%t)V omHEanwȷeGo#[q59V\2n#-cD EX?PNt$l y".]y@n>;!wL)$֌dmuPWbya_[NB 8Td$F@ZT)h [lAZ ʠZ+>wc7n*(ӱ6J6JkJnSn?7i-(רmC]Iݢkt}7%n''V/=ո仙 =OLhPw4YQ㐣|g;pFx;J闁} ~{γ{msϻj91}\^(JJmyw zyOfoX^h| mŎy.IjXfZÐ) 8oj4\AlS(>hujti ޓ5TA! 3ӹZ$䐴[ኒ2к:(;~2)p*|2%Ym 6q٣EOh4>9|{̮*QS+7Ǔz4V _ߞ:w8́caIRَbو"}2wN)&r47'k=gTt7[qu*w~}d=M?Sy4mEiо5rJIfu1_fڶ\n5M^i /sёU + ̫<n<ٚA=$,Džnf 7 P#>}8]n}@N7OqFN< zKePtT r)B*Ar8cE}bQJѵ\Dn,@HnHҁio)$c:&Bб 38FVK )r( 4RKyp2ܘ))`l.wYung&|K/B gf)`]foZiNBx\vogOp7bdoӾog n+AA lv h>K$PzW 4U&WbRpU59%3{/.(Іw\& LPuݯ +/ W!(n6M%8d#kjP Ijܳ|4-S Zڷj ;?Po(Òx*4(vK$Q `ʌHBbVSkA. A㜀O )N0)壒}N1"TE >Rf bo-0,d0@DW!Zqnj:DP;IBREI@zX43(> nam+OC]x7mOWpJ J èX+Rl/BYWxm}<1 -d¶#TihO+Ǒ.m1uPIfJl|p):WRD2*Tw畸_׉PE4ݻ:F;\1bw*\gyſ>~ڞxM\\O}3U.M?DV{WqGtx .ZΣLU!m6, -Ɏ 8\g"~lxdJ mu1bA)*rSvye9C^ttLw@NX*$<4!ۈTTl!gxth'Wv9?~lk2d!# \[ z%,N@[ΜPpJTRZ!).};JP>a9ssc^!ttkO,njG,R+O_cבZ`L gc01HbGMzi6J7YhD?ҕXMZK.6|[D8#Sg_(I- NQ*Lg7T:o 5 TqoϚN+ʘ!29bgTs<). h|KhBzUtcχ ѻ$,W}h^50ͭac6?wQZ-E/_^(|L񏂗jkuR.05zलNi-q(tEVD W?NL(05w ̘eK~YC?A#Ŵc5l?WdJq3~VOs`\zY<ڄܦ[k:6Pz,ɫ>r;:d€6kGFbQdp[翂խ]SĆ&/o'&KȀJ5JQ-0x 8ԸvQ慸;(Q˸ٶREd9(8߆4j.Y L+dl[sJLkX➖xrf+B,^ R(SXE܀uPTBBiU0EwR\WZ]vꑻB LE\s- Aq']V{ub}9*ytl ymvCJ\K/?_ǢQ,&9ٺ@P5.{VDx|rL *oG$^=8KڟK3Y.J-C yfޣ??5pL!8fLq?kqvvY,d@Fl7*ҟ&t]J<|*J>Xx]ЕGb-܁`ɦXJW%M+0њ *^`R`; ?}WWqSVx`@S[]ènpj^\Lzy'u@RFC!m|n0ڝX<ӑVL;;$Bm`eS]{J_?!ʆӍt{.ۗsÛO9S|Yhogy@U2ut~yG~.d(Ӂ2'ЈG~;(G rРED>S]>|3+*է/^3L PHk<)0YYO2}'x>׾8F+<}G?>Dtr~}R)#)<|[ҏ{+ca釆!׷6f*n*XΆ$Uj}<=1X{b #*peX>*Eq X]0 ֹhM8=\"~lxdƵMOIvNIJFtJaz΅C656 ·e<7H#v _b|Dr&PceY]_ !3,3,3,3f r& YI҂E!-|CYw)T[幥A^\zEJ|kn49Ey9͡d͗Ԛp-ӡ0՝1Hjػ6~U2@/ N;=϶H;L{AMvx]9J&9Ȳ# ieƳآB*J9Жõ!@8BD dJA&h d5b|z>jd#vII3yy|0>i&/u`´8D|E>epr;~''#\OzCs R_1,3/% $#Is^zgV儃'˼^Nݘ*YfDs&&OyW<נ8zh5koIǬ6.rRQ!P!EDƅ cm_-8[ VdehEsסe "2P %*(5js l 3p{2by\hIF &&KŨp>e,Pp@B䕵K5ҷTH 3rCIvњVEd|E b.Dืsd[\gED^Ei/ˀ@@z&EкmƭX0ytnb0g| ,Ny~8[5Ons-=[xi `:U{cBaƔC%c:Ӗ@f|ӂ6FaZ(`n;:5B`I O`e&P̟i}{Jl떲b㲘ƪ>9]tQ jVLۍFVn6h, zY-/֗aYOA>aʐQ\,t`fsw6)}" B]s ffu[84&''aXq=יVRIF.h):XF7'_TC2^w4K"i6-vFIE,jOLs(>vc h  A tP7r,6`nX?M;Fv>}{V})U99[obӳ/+/K߇pݡnY\܅EIg|/MmkͲ<ݦE*r!l8uN:tVT)#SưY >ی2+N>@% Xdqd8DguPg'MQ 㚛0ш.s@B$X!TSfRJKxz:~h;_Pe #sUKEFw{c4)<9??k=B-Bj'j4< HI&♥q@+Zn34/i|pxa2҉E_%EFv:%wL̞o7 0|z&a)nMLʮKBb5jfYZ o=E8hDv`0$3ĴEʎѹlEǸn8M5u7d v$&mE͟ϑBZƦN,,LJNK6ሢXA.sp TvF-_ƇLZ$?zƢ|W͑bxT%y8'ZmCtނNK8o?Fp?oԆa#~d:w߻&&Fݐu,#0oJ]CI^rr YB%b |`QJဟv?;':{pn,,o:X&A[կ#Ef ޲|t,:MMdW.PڶJeZL%lI^!4<6ˏxl;LO߿~l9/hب h!!<tqV}*UӪ2 51<0F/q::z~]/6o>]/l#z)lS1vRXJ<>K,Juyw9x5uN3Wk/yٵaJV^$Et0珷a(v0hSTNq4:?Y YNXthhgV_w)ki\jdKS>@թ9(Ub?m95 lNs*T9L c2m= ucxZ9yr}DYEd*.t91FO t&ڢw@^ztzXGət׺rQ 1QY 3u,}HFbiiu=0rhlzs# Lf dŽR,0պ'`ESC?ߕ/@9.ckoz^qZ.U4 g[i% 0))ai@kC B?ןHjƆW':m(PR5t@  q2akJQ^4L"\Ź?4xzsb6Swx,0-:y `LK vYG *iNM-2m5$2J(M,*uupsٙPRQpHᙧ:!$'m$7E/Gd ;sApdOhؖ4l%nk$5N#|*q5b䧔[)ú9)WGFzUͧ몐2G1R- x rꀛ(!a'= 4L} RYnuMSYE Z ckH,SR }ks)ʳB#| LMb!W|x;U:P8"\_Ok[š^]kyx! 4ϑKhu`w;Ś#cLЮZ"Њ׍=rQ[] @#@6XWLHx`ѬVwCJ9d F)Jϓ`vGpTza kH {e5ogb`:@WVRuye hZZJ !J%i25~b1Wlx[-7R9:CG>ulć _Kkp"Qdʼ!'ř:$?l{7w֞Jf1ĉZy2tKKS-{K՚i:pn=YAϽ0uʟ O4V:vջvi/ J,%;S%ǚ?cCfoB7ѾPcs̽k1Ascu,ERO'ٻx{'\J_WaJIh-نJ;1xw'.NG; _~Jp( O 2}W>Ɠto/zJ>Fѯm aob4~vS>}f}'NZ ެ_ԿfZmՑy荘bo~M|sﳍ1bos^/ '*Q5Fw۬`j D!$ 5Q1TظXra@K\jcZ |{Fi >YYiZT:)B&ʦj!чx7([[418Ļ=NDYkܷwk wB&ʦ>kf%;VML;n/E5Zmݺ?>٦af_K?9p\`2-.\/QP đ_DlE!O,u{W~RFPәR Ir0!pu7h .TE˥aM 31 *pİFn-<@ U %ĠߴWŀaP14tB0eQb`շm b¢$8]_</\ Rj{3f~3ET2xnc᳣ϞҼ*ǓPVn_SP`t\1 (#Xɧs/*<Y&Q+4Nxv%O&u3} Ĭ}`|bڳ(1<"]2S`}D  }REbT;,ۆ]U:+?q>s 7]Zb$]ajے0++{ۨl M\y+g_HI% *v v{K-C ŒiUϙp]:wEo,)tʥ̂]WAqh6~ƢGtOg&C8]aĴ`9)M%`Vz+Mq?4C&S)^W$!I]zK(rp=+lB_:`ut/.uɞV4S-%rmd$\ۜ^[5 J`0-a$aOpr{:N``A`/SvZiGbv]S;qp֖.U Kf DEJ/ 3)D#]ِTr^/;IQ y0bI'K?dn_(-%:0c,eV iq ðl}w!!rmZkಅfjv˔efr}g*`*,9%&e+2TުDIբ?~* t|>S*Z5Z% \z;at0Y[y /ٯoGmБ){b7xSݳq`4^z"z7 <:%tcNLeOE-SZ^ݒ\n--)oC ր%w8+Vq.0<[NF8YWR̸"L$]!Zxro*…'e41|*MabGHTM]-m|v+pyw܈ń '&ʣtq࠴(}9Z4{Wz _(enp- ?y@]֒sl]qwӿZP['ê 5~h>n^η/</kԙ._Z%9mj[^@)OE-RZ6Z)V7J1'ZJnʽR stu  Ԏgp5uu BdzunYzN~X.?ǚysoOWCHp'ajѐ̊^J*J+{`?0yHܑ8 i)q Z ֛> l=0w?NA}l1`1[I4V~:0|ӜМ4 Yb*3}!Ŝd;n)fްתI{ژݪ|t$5W$5 [<ΉtpChԋ/6.6;R~H&c}Xqc9`$E!s19 R6l *!g q!7G48vFxx l!efF[\VBD(jAe`c$DLӔ";Tߔ)=Xk-M; (Ot)3±4_ժGf~eZzt3h.4ǫo9`5l/79^ 4 rW4:YaEЪp GGeaC-n˹?UV+!zrquE1XvJ$bw*!H4Xo۟=0Vo ztr)b*1b-M;ByHCYJZG5(mEYGaNPlDryUyxpFKr.E KJ8` t.{;Sn<-^VXtxZ'h9U\P(- &RNcB\yVYRCbVY/UzYHj hRp)ۆKU!6t8寥n{btYvqyqX Пì&~~)ot9O.7,}k&wP`$ KR/~:C.'y_O3Qvzj8%kfkޯ'2Ly{yrr|v;Kp yWZZ8c vOWώxّ Rrw?.hH( }xhn{3; :AYZ;Z^'RbT .Dz!xt UBVՋ^hg y Jn2Gk L0rҳjëYoMף\s=?-+OՆA.K@b/I<}p,y>=a8hK*=2N7oCsPXF@2R82F˕VryF]s6{G x@u=.u8Ab`E- d5D|[]wZnҶ-n5`W⾹{m]>z2*qQjsB+>V]~Jû%m:AӠ+}',[A >xOss24 pΎ2lt>T– r2a'S-l/6V; ƂdJo8+>ٜjTVF.F{ MwS9|bG~o@6ǯ ǚ:fI)O"/O9阞.SePVWf9Aw,1~<=h2ݙPu~6B| ]n}etw{Bɩn/еnCX37^6ew?ݜލܭ Zƞ蔻=*tn/gn+6R ayON|Lay1b,+<k L!EkPApɥVXpE.2zhgx,1d'L#0NqISsˬp4RوA`[|!6 =|.6#=[‘j>3[|r-ځ-BK3l݅؅+ RdvK!%sԥ{|qe~j\it5ǓZڢOojwO Y!ߕ4WhU;:կ ,-W;bm]XCJ$k3ge!#έp;Dp~ U|^ B*b,Ӡ2 }M0h.uJ:(+6Տ0{'W,0U6i&_<:3Jdco\N\ۥ8LjH0N`X?cC" D-Zg RR 2y=w; 䬃j7^ Z"OED/LBO`(YxFœ0ല>Rmzl-~nQð"YT=$$%ٱ&kB1)]J)exǧt 5yGv@?ѿV4Yl`Vs#/ AsA\\ !7tR ejJe_,b_*괯,\n1p@h,%sz` "%BYa(cQIޓJ L (7%>VqÿsA 1;H݋ F[ ͋ H$}j 8hQ҈ "`-B:/%89$Дdy%Zz~MK~͈;< l [ ]!h9x>%Sv]@w_ rv_n5﷙þjr%䠮WeAJ`"jo9U@`g-$e']'R}[3G8|ߗ1P;NҩB0"ƣ02UaQAp Y֍ԗ{ɥ-qNGlz?> s+sfXK OBgTW5 縳*hϭZosxǔK@nKõoOYZl nNj`Q}C^܇C-J4 1UsPd UVlkƶ%vAT׳=f4.+sѫf-;W@PB=Vx.& dM-Yu*! S?|_'LkXekJ7ˋTgQl0G'>EBŴR<3S,+R?~M>4Z4` K!η"Fg ޢ0D . L}0l 5h@n8@΀DXM]s'ݿPx[uekIYCihHC!P< +!B%! یҰ#Y50JR &mELDgz,B|` y+ 8-S2']fg QSy7mFy+oƐ{X2(s&tzT|`9i1ĉt^EL..҆{`L[0ȝ3I/iI*%jZ.,J"{o3ZS#i-Z8͛Z,?s'cFW pTz|9\MwS*7=ŋf[㻩?L{ZqxWݕ"uwӊI{xOstO&GgΦM4K%+[\TҨy@l4c[%۽bU)0>Ԍjf;KƾU;Gi2`y#rW=}'D剟.g1z=7)91[%}`y|~Nڏ⸗gr8L'n+CFD˴G5x dP7µTC뉑m&F}2ύVTͽU=֩/9^Ve30ΛY/z[/|7$N!/ )h o(!>'uB*was-όC޾zc4 ̫j+f`UvXi¦^;bɱ)ͱ=;vYlns;޼6wKl9Mkg-4w\aպx˸0wi׳U:n#cl5R|>N'5ji"(_s%&M^{'orzK!R8Vhy vdaqZ4DzIw~?&rWQ3gAZ1=6L@T s?6e"CQ,́J-]CĻ6#ܒxʖ{ RBS[K"TVuӅ<۸7yu4x+ L NZ {Ljɼ#h!LPb6I!1XBa9Yq^lN+NTFR9; ,@S EOqQZhUR>T#!PxDBF5C"Yb)EGkTMCa']Iz$`abd̪:db%P!dIț=;Q2@P.`xi]@@qgJ/h! 2a L{k} TΔbFYeCĸQՎz82tU/$1h8($PZF+##'i5dG?d|(N"xsz9#(rr|rt⇕n+ݷ- 6@d?{G_ e>dр$$>a,xe;=Ş3hlI 5,EjD jѼ}٦#_jz܂ʉ؏Wz8^urO#q{A\=ڌ$B^s~8$f9wKJt> oŎ{$5oؒ2C4خ8Rc?ZŘ3Y7Eqqyl-38EJe4!mel!*oSjne7 Z: ݢ 51v=}JP6vޟ/Ҿȏ" V霥IMj+cZl.GکI_;ܯkd{5/i=~gOs8կ;cUϗF3oy&Vo#/^m#׹6rõKH?k!s$:0=޹pS`9r* P#GYƠ6J]jsQ&\%膽Ǥts&|DZ?yyUwYGmٔbx.c9}ym.ȭټx<]Nj2I\8'Wo63?ii313nO)/g9&Z58UM_^=s/J):Mw]O ,EO&KYRb ˥ѐ߹N)|&>uuSmtm*jTggn(?ߕmYuk!sS\Tg\ZdG܃ 3*+cla<چ;OGu~z>a'fH ;Nhf =?Qn-:}\VǙ`a/K3hotWr S/.>oS s#j~cF4j0#Q~2fv4vKqEY7l;2xw5l5k&5[5Ӊ7lxUS&KʧVt^C褧;B:y -ִs %8x KO) & -ּqC3,"Σt:C5ohB91jrRgfkܮ5oF$2aAyZ/;PZ<DQ$| :{<OlڱC}jpX!*01e(r~q旇$KEH}H91?-b/GVtSo]F=y=6_Kdz-E`i}iA bqޟ4oX|};9X'i ʋ?_q_r׷Z[.ԓw]Zo_]3V,{H4d3w"i Ē%w.;,y{߼td,d^=heQw?:`&qr"BeMrpl i5= eS:RQR2lI IL@`y,D#9yH3$Sa2b1d72zt2!E@6P+VZaU|t>GWq=lY} 3, lCrr,dM*Ͱ\IIc5(W],|Ȯtom"Z׊Gr"̝:1 '7l#Tof{RܫU/_;^Q>]v+1RJ1 P0gwKWvDJQVmeR(&6V=>EAUh`VK4Eu! E8]3;9k8Ij>Pނ+fWuBC&wPn,}L[b@&Ny wOi3y ` 9)?@#[V/8 T xqLOBax_(g@13O õ82ie63 Gh2,&C3mgvr1&Vßt@ l{?q;D#M&|z ɥ qrnCpngptf5DŽgpP]l_׆x&uׁʪˎ1axv(@D5r{yĶdԉmGw^'w|hB3W~Y*? nq;O6Mw$X춣G5%rvPhzIL0^bWl/О25*%i❚ =km#H6Z=׳lJ,hRR -9%WYnJ"ʄZ=͔Dhi؟W)7ʙBTSZ!Z^vjY_SH Nm-7BK$@j |%y?JPsKtoDZUO=?!-k+Vvٷ+rшNe=| D wRJ#~[ Ϗ%bB'w; |0pD`vRƬ2FF$[mO 9 r )F2b{Nz6Q.؋ 6Tq(-7Lz\D@lDV)J*Ps2)3YQUQQZWGeE x{HL@iA,Ϋ}Иr jJi9GBMM @l?/S@*n4̠L\6kcXef7VZ)dS5a(}h5v:(]0XE)-{Ҵ&>ZFo5KdDcP(_&[^x { .!1R+L@.& 7 wH N2uG\獇EŽP-īʶ[eN4,T[,կ[j jJF쾠xPYu鸀m^!~@LQzQ;gZƋ܁"cl][)IItκD|<5oF|-DJ=jCAm__W^)y3҇QDqcߣ ^|,!Q?*7.Zw"Y`QH,CAKn?Ɓ$XPЙ 2Ϭ.Rb ɚ7H7R/ $x^~Z%=.Q`uZ })DyBDzM7'ZOλqVx FXg7 u|c [UUxŋA8?ꜟZ2sv~X9 3ϵRfy?ہӁp̞P:lԆm5X+ikgUOVs#0ZjT|cxuHϿGYK2Z#0T%[ak-Hcs[.R܋eB|0x{Vi/3"|lEh}6Cʑڈty?0 \F\tԎ:Q4YЁf̒S3M1K lw<9'-ƚ3'w9A=ua5X L:WLS~QVO'HBq/%<864$F ef^Cɚ  9F+dNxC" #*H rBc"-ytI4D\1C幹ԃjt`|KT dD 9*2&L,|D y+ $J`CٹGTD*Qn#_X’Vݺ۟嬁N+?oʯsXUbߐ1 V~'*BSc邕߉9F;"58G.ݘ`RCNUR\*[g.]vٝ'֍\7"'qpL=Hḱ.\/׼act=T: 1yҟY#?z֔iZ 'eWO?O#:D,?wL luv73 ڴ4+{XQ–qsu jKnk2lvWB/ta,}X|qh6F͌t3I$oJGtbw9k%*jac`bbBWonOJP[+lAN.&O'jd[jpRy5l^_dP+:j+<̓cc5ol#eue.'`e=Ap4.N2A{@5dh߇zz6D4tzW ~5Xӭd^oh!TYE-X Q#Z"Jl7HLJ65%:JR7VIz^)Ydv9Dh袌)8tT:ي Fw+udRVKǃFfV )~+ʹ}x9ЊzkZi_>\P}8lc vր3Fͼa,2=qpf[L<͌[,,!]%- L͆ Hiv6zg$c@Z[u^, Tr=x&9o`߬2 ZI$}^msFzgs e`czeN(\i$[F;i =5uRVOiicrpGgB\L? Qa;W1eBi'hmQ[*=; `!mS@\nZ2(D*!3SYCɕ&+Y+r2٠Eg5BqY}2qlY㾂B;䡛;_1 *].{M1KjGMP9ѱ:%񠪈(N+%Ծ6 jY`o{"+:&Ql l?>C6C05W2&2@+].l)7㖇 lY쫛<@j mۧI/c|?ݽ51y~;|fI=Yj7e)דNj(Pτӿ_GZӓ~GϭjxG_֚E Sۭz}ee' ~R 'oY3~WY؆9 eey>m Q4?"|wRA9 ǽugױBì8'> B-wNȁT^9N P'~<8 ,F."Kt4o&ƹtTFDلՒ|N&lzpUQgapsrkMW)7׍d-[;gŲkXIR,ﹻe1n}vXz%͇ Ί_\7/,tR/(^؉x-'{ү>$yk a5YAeचf@E{6هOg.t>ifSJ#K0I.Oшt9lWʌ%д 8d4wR_@A@CXRg QEK8 Jc8Sf346(B5o؂"eH@z'99lSkts!]^-pvV ?Y{ZץoG kNgоeVHAKLOE^οjp>f?/W.+sՠVv΅mn,SewZv;Gہh[D'wXIdX;%\w۴u)hap Ynt,!K؊P}-*u~J]6T~ ˊF"-`$ꝸn]d<`靫f2FX"Q́95U\-HjtAœ֍y2CTZLm5FYKOMy y];[-MTvH H==sakr H8vNZi0ExI˝ ~zCVz+\62nT#p5.:3ޜ%ӞI+U$;suXS#M_w1=!wyaHng=?@K;rus mfݥڻHK%'2W,;f22w(.[a̻TYziR\uLvceA2D&\=4`9֐tb>غc}5o++I,1?:6G7uNiyT84CBsR$&!w%Ic3Z֊~TZZZVR-Q]FTmQ2AܙzwkTkA3E&Jr.1R;ii& t@n֣}̍`9.I/ N}!0jo!D.$ۖb/BAPt"E))jhG6?$Evf 2hї)ZIe>۲D2ɘt#R{HXvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000002622132615157273421017710 0ustar rootrootMar 20 16:00:47 crc systemd[1]: Starting Kubernetes Kubelet... Mar 20 16:00:47 crc restorecon[4681]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:47 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 16:00:48 crc restorecon[4681]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 20 16:00:49 crc kubenswrapper[4936]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 16:00:49 crc kubenswrapper[4936]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 20 16:00:49 crc kubenswrapper[4936]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 16:00:49 crc kubenswrapper[4936]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 16:00:49 crc kubenswrapper[4936]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 20 16:00:49 crc kubenswrapper[4936]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.563711 4936 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570573 4936 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570604 4936 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570615 4936 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570625 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570633 4936 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570641 4936 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570665 4936 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570673 4936 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570682 4936 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570690 4936 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570697 4936 feature_gate.go:330] unrecognized feature gate: Example Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570705 4936 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570712 4936 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570720 4936 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570728 4936 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570736 4936 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570743 4936 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570751 4936 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570759 4936 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570766 4936 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570773 4936 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570781 4936 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570788 4936 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570798 4936 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570806 4936 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570813 4936 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570821 4936 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570828 4936 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570836 4936 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570844 4936 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570852 4936 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570859 4936 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570869 4936 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570880 4936 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570889 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570898 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570907 4936 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570917 4936 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570929 4936 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570941 4936 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570952 4936 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570961 4936 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570969 4936 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570977 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570985 4936 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.570993 4936 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571000 4936 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571008 4936 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571015 4936 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571023 4936 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571030 4936 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571038 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571045 4936 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571055 4936 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571062 4936 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571070 4936 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571077 4936 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571084 4936 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571092 4936 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571100 4936 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571108 4936 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571118 4936 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571128 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571137 4936 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571145 4936 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571152 4936 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571160 4936 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571167 4936 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571175 4936 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571182 4936 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.571190 4936 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571349 4936 flags.go:64] FLAG: --address="0.0.0.0" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571367 4936 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571381 4936 flags.go:64] FLAG: --anonymous-auth="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571393 4936 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571405 4936 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571415 4936 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571427 4936 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571438 4936 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571447 4936 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571456 4936 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571466 4936 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571476 4936 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571485 4936 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571494 4936 flags.go:64] FLAG: --cgroup-root="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571502 4936 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571511 4936 flags.go:64] FLAG: --client-ca-file="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571520 4936 flags.go:64] FLAG: --cloud-config="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571528 4936 flags.go:64] FLAG: --cloud-provider="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571536 4936 flags.go:64] FLAG: --cluster-dns="[]" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571570 4936 flags.go:64] FLAG: --cluster-domain="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571579 4936 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571588 4936 flags.go:64] FLAG: --config-dir="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571597 4936 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571606 4936 flags.go:64] FLAG: --container-log-max-files="5" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571619 4936 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571628 4936 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571637 4936 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571646 4936 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571655 4936 flags.go:64] FLAG: --contention-profiling="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571664 4936 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571673 4936 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571685 4936 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571694 4936 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571704 4936 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571713 4936 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571722 4936 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571731 4936 flags.go:64] FLAG: --enable-load-reader="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571740 4936 flags.go:64] FLAG: --enable-server="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571749 4936 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571760 4936 flags.go:64] FLAG: --event-burst="100" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571769 4936 flags.go:64] FLAG: --event-qps="50" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571779 4936 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571788 4936 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571797 4936 flags.go:64] FLAG: --eviction-hard="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571807 4936 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571816 4936 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571825 4936 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571834 4936 flags.go:64] FLAG: --eviction-soft="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571843 4936 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571852 4936 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571861 4936 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571870 4936 flags.go:64] FLAG: --experimental-mounter-path="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571879 4936 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571888 4936 flags.go:64] FLAG: --fail-swap-on="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571897 4936 flags.go:64] FLAG: --feature-gates="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571907 4936 flags.go:64] FLAG: --file-check-frequency="20s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571916 4936 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571925 4936 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571934 4936 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571943 4936 flags.go:64] FLAG: --healthz-port="10248" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571953 4936 flags.go:64] FLAG: --help="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571962 4936 flags.go:64] FLAG: --hostname-override="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571971 4936 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571980 4936 flags.go:64] FLAG: --http-check-frequency="20s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571989 4936 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.571997 4936 flags.go:64] FLAG: --image-credential-provider-config="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572006 4936 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572015 4936 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572023 4936 flags.go:64] FLAG: --image-service-endpoint="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572032 4936 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572040 4936 flags.go:64] FLAG: --kube-api-burst="100" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572049 4936 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572059 4936 flags.go:64] FLAG: --kube-api-qps="50" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572067 4936 flags.go:64] FLAG: --kube-reserved="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572076 4936 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572084 4936 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572093 4936 flags.go:64] FLAG: --kubelet-cgroups="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572102 4936 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572111 4936 flags.go:64] FLAG: --lock-file="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572120 4936 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572129 4936 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572138 4936 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572151 4936 flags.go:64] FLAG: --log-json-split-stream="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572160 4936 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572169 4936 flags.go:64] FLAG: --log-text-split-stream="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572178 4936 flags.go:64] FLAG: --logging-format="text" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572187 4936 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572196 4936 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572205 4936 flags.go:64] FLAG: --manifest-url="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572214 4936 flags.go:64] FLAG: --manifest-url-header="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572225 4936 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572234 4936 flags.go:64] FLAG: --max-open-files="1000000" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572245 4936 flags.go:64] FLAG: --max-pods="110" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572254 4936 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572263 4936 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572272 4936 flags.go:64] FLAG: --memory-manager-policy="None" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572282 4936 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572291 4936 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572300 4936 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572309 4936 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572330 4936 flags.go:64] FLAG: --node-status-max-images="50" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572339 4936 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572348 4936 flags.go:64] FLAG: --oom-score-adj="-999" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572357 4936 flags.go:64] FLAG: --pod-cidr="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572366 4936 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572379 4936 flags.go:64] FLAG: --pod-manifest-path="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572388 4936 flags.go:64] FLAG: --pod-max-pids="-1" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572397 4936 flags.go:64] FLAG: --pods-per-core="0" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572406 4936 flags.go:64] FLAG: --port="10250" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572415 4936 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572424 4936 flags.go:64] FLAG: --provider-id="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572432 4936 flags.go:64] FLAG: --qos-reserved="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572441 4936 flags.go:64] FLAG: --read-only-port="10255" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572459 4936 flags.go:64] FLAG: --register-node="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572469 4936 flags.go:64] FLAG: --register-schedulable="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572477 4936 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572491 4936 flags.go:64] FLAG: --registry-burst="10" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572500 4936 flags.go:64] FLAG: --registry-qps="5" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572509 4936 flags.go:64] FLAG: --reserved-cpus="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572518 4936 flags.go:64] FLAG: --reserved-memory="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572528 4936 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572558 4936 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572567 4936 flags.go:64] FLAG: --rotate-certificates="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572576 4936 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572585 4936 flags.go:64] FLAG: --runonce="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572594 4936 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572603 4936 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572612 4936 flags.go:64] FLAG: --seccomp-default="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572621 4936 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572630 4936 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572640 4936 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572649 4936 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572659 4936 flags.go:64] FLAG: --storage-driver-password="root" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572668 4936 flags.go:64] FLAG: --storage-driver-secure="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572678 4936 flags.go:64] FLAG: --storage-driver-table="stats" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572686 4936 flags.go:64] FLAG: --storage-driver-user="root" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572695 4936 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572704 4936 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572712 4936 flags.go:64] FLAG: --system-cgroups="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572721 4936 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572735 4936 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572743 4936 flags.go:64] FLAG: --tls-cert-file="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572751 4936 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572761 4936 flags.go:64] FLAG: --tls-min-version="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572770 4936 flags.go:64] FLAG: --tls-private-key-file="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572782 4936 flags.go:64] FLAG: --topology-manager-policy="none" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572791 4936 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572799 4936 flags.go:64] FLAG: --topology-manager-scope="container" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572808 4936 flags.go:64] FLAG: --v="2" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572821 4936 flags.go:64] FLAG: --version="false" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572832 4936 flags.go:64] FLAG: --vmodule="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572843 4936 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.572852 4936 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573080 4936 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573091 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573101 4936 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573109 4936 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573118 4936 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573128 4936 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573138 4936 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573147 4936 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573157 4936 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573165 4936 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573173 4936 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573181 4936 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573189 4936 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573197 4936 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573205 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573213 4936 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573221 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573254 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573264 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573272 4936 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573280 4936 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573288 4936 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573295 4936 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573303 4936 feature_gate.go:330] unrecognized feature gate: Example Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573314 4936 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573323 4936 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573330 4936 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573337 4936 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573345 4936 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573352 4936 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573360 4936 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573368 4936 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573378 4936 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573388 4936 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573397 4936 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573406 4936 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573415 4936 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573429 4936 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573440 4936 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573450 4936 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573460 4936 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573469 4936 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573477 4936 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573484 4936 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573493 4936 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573501 4936 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573509 4936 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573517 4936 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573525 4936 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573533 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573541 4936 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573575 4936 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573584 4936 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573591 4936 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573599 4936 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573607 4936 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573618 4936 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573625 4936 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573633 4936 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573641 4936 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573649 4936 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573657 4936 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573664 4936 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573672 4936 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573679 4936 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573687 4936 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573695 4936 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573702 4936 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573710 4936 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573720 4936 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.573728 4936 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.573751 4936 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.589391 4936 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.589459 4936 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589680 4936 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589697 4936 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589707 4936 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589716 4936 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589726 4936 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589736 4936 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589745 4936 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589754 4936 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589762 4936 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589770 4936 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589780 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589790 4936 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589801 4936 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589811 4936 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589821 4936 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589831 4936 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589842 4936 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589852 4936 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589862 4936 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589871 4936 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589881 4936 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589891 4936 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589898 4936 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589906 4936 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589914 4936 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589922 4936 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589930 4936 feature_gate.go:330] unrecognized feature gate: Example Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589938 4936 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589946 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589953 4936 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589961 4936 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589969 4936 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589979 4936 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589987 4936 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.589995 4936 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590003 4936 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590011 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590020 4936 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590028 4936 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590037 4936 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590045 4936 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590057 4936 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590069 4936 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590079 4936 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590087 4936 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590098 4936 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590112 4936 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590120 4936 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590129 4936 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590138 4936 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590149 4936 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590160 4936 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590169 4936 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590177 4936 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590185 4936 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590194 4936 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590202 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590211 4936 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590219 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590227 4936 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590236 4936 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590245 4936 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590255 4936 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590262 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590270 4936 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590278 4936 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590288 4936 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590298 4936 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590306 4936 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590315 4936 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590339 4936 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.590354 4936 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590652 4936 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590671 4936 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590681 4936 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590689 4936 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590697 4936 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590705 4936 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590713 4936 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590721 4936 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590729 4936 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590737 4936 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590745 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590753 4936 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590761 4936 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590769 4936 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590776 4936 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590785 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590792 4936 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590800 4936 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590808 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590815 4936 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590825 4936 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590833 4936 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590841 4936 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590848 4936 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590856 4936 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590864 4936 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590872 4936 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590880 4936 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590888 4936 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590898 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590908 4936 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590922 4936 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590935 4936 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590945 4936 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590954 4936 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590964 4936 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590974 4936 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590989 4936 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.590999 4936 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591009 4936 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591019 4936 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591034 4936 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591046 4936 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591057 4936 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591068 4936 feature_gate.go:330] unrecognized feature gate: Example Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591078 4936 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591089 4936 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591099 4936 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591109 4936 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591119 4936 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591128 4936 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591138 4936 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591148 4936 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591156 4936 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591164 4936 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591172 4936 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591179 4936 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591187 4936 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591195 4936 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591203 4936 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591211 4936 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591219 4936 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591230 4936 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591239 4936 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591249 4936 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591256 4936 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591264 4936 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591272 4936 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591283 4936 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591293 4936 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.591302 4936 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.591314 4936 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.591652 4936 server.go:940] "Client rotation is on, will bootstrap in background" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.598723 4936 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.604134 4936 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.604319 4936 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.606569 4936 server.go:997] "Starting client certificate rotation" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.606617 4936 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.606837 4936 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.635272 4936 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.638197 4936 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.639107 4936 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.656737 4936 log.go:25] "Validated CRI v1 runtime API" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.699993 4936 log.go:25] "Validated CRI v1 image API" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.702990 4936 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.708878 4936 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-20-15-56-05-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.708936 4936 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.744083 4936 manager.go:217] Machine: {Timestamp:2026-03-20 16:00:49.741330623 +0000 UTC m=+0.687698508 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f03975ca-8846-4c96-a4d2-0fc5b8167315 BootID:4fc88e84-3312-4795-974e-5b099fa998eb Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:61:3f:e0 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:61:3f:e0 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:4d:78:36 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:35:da:f5 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:30:d4:13 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d7:49:cc Speed:-1 Mtu:1496} {Name:eth10 MacAddress:12:2f:47:b2:c9:af Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:5e:47:f5:d9:6f:35 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.744638 4936 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.744936 4936 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.747686 4936 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.748125 4936 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.748180 4936 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.748533 4936 topology_manager.go:138] "Creating topology manager with none policy" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.748596 4936 container_manager_linux.go:303] "Creating device plugin manager" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.749194 4936 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.749305 4936 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.749836 4936 state_mem.go:36] "Initialized new in-memory state store" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.749987 4936 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.754592 4936 kubelet.go:418] "Attempting to sync node with API server" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.754634 4936 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.754678 4936 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.754704 4936 kubelet.go:324] "Adding apiserver pod source" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.754772 4936 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.760885 4936 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.762049 4936 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.762702 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.762731 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.762834 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.762865 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.766042 4936 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.767969 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768015 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768035 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768050 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768074 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768089 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768104 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768128 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768144 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768160 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768209 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768226 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768271 4936 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.768959 4936 server.go:1280] "Started kubelet" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.769345 4936 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.769322 4936 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.770529 4936 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.771058 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:49 crc systemd[1]: Started Kubernetes Kubelet. Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.774521 4936 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.774619 4936 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.775031 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.775073 4936 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.775105 4936 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.775297 4936 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.775877 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="200ms" Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.776025 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.779805 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.781127 4936 factory.go:55] Registering systemd factory Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.781176 4936 factory.go:221] Registration of the systemd container factory successfully Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.781830 4936 server.go:460] "Adding debug handlers to kubelet server" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.779079 4936 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.110:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189e9804d60b5fe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,LastTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.785633 4936 factory.go:153] Registering CRI-O factory Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.785700 4936 factory.go:221] Registration of the crio container factory successfully Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.785841 4936 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.785879 4936 factory.go:103] Registering Raw factory Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.785907 4936 manager.go:1196] Started watching for new ooms in manager Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.790265 4936 manager.go:319] Starting recovery of all containers Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800655 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800740 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800767 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800789 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800811 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800830 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800850 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800871 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800894 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800916 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800936 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800955 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800974 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.800997 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.801016 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.801037 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.801072 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.801091 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.801112 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.801140 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.801167 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.801196 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807169 4936 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807248 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807283 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807319 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807353 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807465 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807497 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807524 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807619 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807656 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807681 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807706 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807730 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807757 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807784 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807812 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807836 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807861 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807888 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807911 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807935 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807961 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.807987 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808015 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808043 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808078 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808104 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808134 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808160 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808188 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808217 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808258 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808289 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808318 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808346 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808375 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808406 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808433 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808463 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808489 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808514 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808577 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808607 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808634 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808658 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808682 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808706 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808734 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808764 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808787 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808812 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808840 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808867 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808894 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808922 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808949 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.808978 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809009 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809039 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809066 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809096 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809121 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809148 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809182 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809211 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809268 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809296 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809322 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809348 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809375 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809403 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809432 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809457 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809488 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809517 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809580 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809613 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809646 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809674 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809708 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809739 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809768 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809797 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809842 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809889 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809919 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809952 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809981 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.809147 4936 manager.go:324] Recovery completed Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810013 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810155 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810177 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810194 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810209 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810224 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810238 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810252 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810265 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810287 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810301 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810313 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810325 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810337 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810349 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810360 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810405 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810418 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810430 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810442 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810454 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810467 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810479 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810490 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810502 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810514 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810528 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810556 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810568 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810581 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810593 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810604 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810616 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810630 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810641 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810652 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810664 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810677 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810689 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810700 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810713 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810724 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810737 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810747 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810758 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810773 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810786 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810797 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810809 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810821 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810831 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810841 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810852 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810864 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810882 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810896 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810908 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810920 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810931 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810943 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810957 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810969 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810982 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.810994 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811005 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811018 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811031 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811044 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811056 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811067 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811079 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811092 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811103 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811115 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811129 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811140 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811174 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811185 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811197 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811210 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811222 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811235 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811247 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811262 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811274 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811286 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811302 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811316 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811327 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811340 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811352 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811363 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811375 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811386 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811397 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811410 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811422 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811433 4936 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811445 4936 reconstruct.go:97] "Volume reconstruction finished" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.811454 4936 reconciler.go:26] "Reconciler: start to sync state" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.826649 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.829111 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.829161 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.829192 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.833746 4936 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.833779 4936 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.833806 4936 state_mem.go:36] "Initialized new in-memory state store" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.849923 4936 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.851641 4936 policy_none.go:49] "None policy: Start" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.852835 4936 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.852889 4936 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.852921 4936 kubelet.go:2335] "Starting kubelet main sync loop" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.852971 4936 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.852985 4936 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.853016 4936 state_mem.go:35] "Initializing new in-memory state store" Mar 20 16:00:49 crc kubenswrapper[4936]: W0320 16:00:49.854276 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.854357 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.875889 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.904739 4936 manager.go:334] "Starting Device Plugin manager" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.904806 4936 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.904820 4936 server.go:79] "Starting device plugin registration server" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.905467 4936 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.905484 4936 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.905768 4936 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.906029 4936 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.906301 4936 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.915804 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.954662 4936 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.954824 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.956415 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.956456 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.956484 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.956667 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.957025 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.957088 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.957809 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.957893 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.957922 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.958209 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.958318 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.958365 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.958209 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.958656 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.958672 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.959593 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.959638 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.959656 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.960843 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.960890 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.960911 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.961096 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.961209 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.961269 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.962247 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.962279 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.962294 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.962301 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.962305 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.962316 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.962472 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.962608 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.962678 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.963258 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.963310 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.963334 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.963611 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.963675 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.963898 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.963922 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.963934 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.964786 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.964811 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:49 crc kubenswrapper[4936]: I0320 16:00:49.964821 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:49 crc kubenswrapper[4936]: E0320 16:00:49.977182 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="400ms" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.005928 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.007205 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.007253 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.007263 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.007292 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:00:50 crc kubenswrapper[4936]: E0320 16:00:50.007852 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.110:6443: connect: connection refused" node="crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015347 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015391 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015444 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015471 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015525 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015569 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015588 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015604 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015642 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015672 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015693 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015714 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015737 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015762 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.015788 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116456 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116614 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116651 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116684 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116716 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116744 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116775 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116809 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116841 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116833 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116937 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.116872 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117005 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117017 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117047 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117066 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117078 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117110 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117116 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117145 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117166 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117212 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117299 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117345 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117390 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117433 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117477 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117514 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117587 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.117627 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.208836 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.210795 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.210859 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.210878 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.210914 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:00:50 crc kubenswrapper[4936]: E0320 16:00:50.211671 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.110:6443: connect: connection refused" node="crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.308283 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.323027 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.344186 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.367412 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: W0320 16:00:50.367445 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-f66dac0c5b6de71cddf46285232791e42faf202215da962f07bc20832a3ad1e0 WatchSource:0}: Error finding container f66dac0c5b6de71cddf46285232791e42faf202215da962f07bc20832a3ad1e0: Status 404 returned error can't find the container with id f66dac0c5b6de71cddf46285232791e42faf202215da962f07bc20832a3ad1e0 Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.376281 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:50 crc kubenswrapper[4936]: E0320 16:00:50.378022 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="800ms" Mar 20 16:00:50 crc kubenswrapper[4936]: W0320 16:00:50.383016 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-e4830a3386a362b06be1e3743234e51e2800799de923217e7cc7eec426508504 WatchSource:0}: Error finding container e4830a3386a362b06be1e3743234e51e2800799de923217e7cc7eec426508504: Status 404 returned error can't find the container with id e4830a3386a362b06be1e3743234e51e2800799de923217e7cc7eec426508504 Mar 20 16:00:50 crc kubenswrapper[4936]: W0320 16:00:50.389158 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-cc688fe436386973af30cb7367d0efb9c9f3579e90081cff66d744ec40070549 WatchSource:0}: Error finding container cc688fe436386973af30cb7367d0efb9c9f3579e90081cff66d744ec40070549: Status 404 returned error can't find the container with id cc688fe436386973af30cb7367d0efb9c9f3579e90081cff66d744ec40070549 Mar 20 16:00:50 crc kubenswrapper[4936]: W0320 16:00:50.406641 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-bfbdeb2acb05f7d71a0c769e97316d2ae149acc11c3184cb15bf729c440ffefd WatchSource:0}: Error finding container bfbdeb2acb05f7d71a0c769e97316d2ae149acc11c3184cb15bf729c440ffefd: Status 404 returned error can't find the container with id bfbdeb2acb05f7d71a0c769e97316d2ae149acc11c3184cb15bf729c440ffefd Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.612518 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.614811 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.614878 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.614899 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.614937 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:00:50 crc kubenswrapper[4936]: E0320 16:00:50.615637 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.110:6443: connect: connection refused" node="crc" Mar 20 16:00:50 crc kubenswrapper[4936]: W0320 16:00:50.622435 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:50 crc kubenswrapper[4936]: E0320 16:00:50.622603 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.773181 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.863060 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"605124a4518926d695cef4dcd7b0091564fb7ce8e4b191f164aa2884064860db"} Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.864482 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bfbdeb2acb05f7d71a0c769e97316d2ae149acc11c3184cb15bf729c440ffefd"} Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.866424 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cc688fe436386973af30cb7367d0efb9c9f3579e90081cff66d744ec40070549"} Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.867943 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e4830a3386a362b06be1e3743234e51e2800799de923217e7cc7eec426508504"} Mar 20 16:00:50 crc kubenswrapper[4936]: I0320 16:00:50.869321 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f66dac0c5b6de71cddf46285232791e42faf202215da962f07bc20832a3ad1e0"} Mar 20 16:00:50 crc kubenswrapper[4936]: W0320 16:00:50.887785 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:50 crc kubenswrapper[4936]: E0320 16:00:50.887881 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:51 crc kubenswrapper[4936]: W0320 16:00:51.017830 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:51 crc kubenswrapper[4936]: E0320 16:00:51.017930 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:51 crc kubenswrapper[4936]: W0320 16:00:51.170267 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:51 crc kubenswrapper[4936]: E0320 16:00:51.170453 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:51 crc kubenswrapper[4936]: E0320 16:00:51.179811 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="1.6s" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.416562 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.418319 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.418379 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.418392 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.418432 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:00:51 crc kubenswrapper[4936]: E0320 16:00:51.419243 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.110:6443: connect: connection refused" node="crc" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.701471 4936 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 16:00:51 crc kubenswrapper[4936]: E0320 16:00:51.702936 4936 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.772439 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.878332 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"161e8a105db0ed826e7d924f9a1556199b4812d8f0c5190581f2286b23c78aa7"} Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.878441 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c1efc7a1f0b98ede844eee4ce096aa707924942a4460af4ac095aef8f0c9acc3"} Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.881098 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133" exitCode=0 Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.881179 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133"} Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.881356 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.883605 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.883668 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.883694 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.885846 4936 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="843f06f26ea027d49e7dfb0f54841e971808cdf58b4d917221f3883df50568da" exitCode=0 Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.885955 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"843f06f26ea027d49e7dfb0f54841e971808cdf58b4d917221f3883df50568da"} Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.886086 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.886125 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.887714 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.887766 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.887786 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.887837 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.887882 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.887903 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.889449 4936 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6223625140bc1b509d454745a703ca77786e327e82f128961287db69e9197f62" exitCode=0 Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.889584 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6223625140bc1b509d454745a703ca77786e327e82f128961287db69e9197f62"} Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.889775 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.892348 4936 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f481c0dfdc9c5c245f947fd832a2ef774bcb1ddb9a9c2d582312701ce99f5f17" exitCode=0 Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.892388 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f481c0dfdc9c5c245f947fd832a2ef774bcb1ddb9a9c2d582312701ce99f5f17"} Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.892581 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.892834 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.892890 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.892913 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.893958 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.894014 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:51 crc kubenswrapper[4936]: I0320 16:00:51.894036 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.772391 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:52 crc kubenswrapper[4936]: E0320 16:00:52.781650 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="3.2s" Mar 20 16:00:52 crc kubenswrapper[4936]: W0320 16:00:52.810302 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:52 crc kubenswrapper[4936]: E0320 16:00:52.810437 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.896467 4936 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="75df1ca6583b09037c3515a2a96e39b1f203738c70abe1227dbc22c257eff6c7" exitCode=0 Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.896624 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"75df1ca6583b09037c3515a2a96e39b1f203738c70abe1227dbc22c257eff6c7"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.896705 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.898528 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.898595 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.898607 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.900727 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"00b64b5454516549eadf231e70c771f6306ba4ab55c2ea781752ec6516c8213e"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.900787 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.901841 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.901872 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.901887 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.904571 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ad92eb888719c742dcf914af04fdd94261c36e1dea7440c0845498bbb331b370"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.904634 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e18a07712a45a16e535f7511f9160910b64950f2bcbae6d551c11a75095e7fac"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.904658 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"370c00a9c484dca5d070fe093430cb7eb8ce680ee723a1574237297fdf5d41e7"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.904637 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.906149 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.906204 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.906228 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.908489 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3910aad0bf2f4b2dddd0762a0e09ef9f8d4de384902ab2074ef8a4fc5529b721"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.908537 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"73cde5934363a3a504adcdf396e9fe89e6750a2128ff18681bb6fed40c7e9ac6"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.908610 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.909731 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.909780 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.909800 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.912846 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.912886 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.912904 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8"} Mar 20 16:00:52 crc kubenswrapper[4936]: I0320 16:00:52.912921 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929"} Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.020458 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:53 crc kubenswrapper[4936]: W0320 16:00:53.021152 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:53 crc kubenswrapper[4936]: E0320 16:00:53.021296 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.024818 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.024936 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.024954 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.025023 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:00:53 crc kubenswrapper[4936]: E0320 16:00:53.025901 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.110:6443: connect: connection refused" node="crc" Mar 20 16:00:53 crc kubenswrapper[4936]: W0320 16:00:53.028400 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:53 crc kubenswrapper[4936]: E0320 16:00:53.028470 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.226330 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.235216 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:53 crc kubenswrapper[4936]: W0320 16:00:53.255296 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.110:6443: connect: connection refused Mar 20 16:00:53 crc kubenswrapper[4936]: E0320 16:00:53.255574 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.110:6443: connect: connection refused" logger="UnhandledError" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.838442 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.922609 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7ad9886cd7aa3c83e1c258952be05467e5d16f14b8701041a35cd852c4c07489"} Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.923066 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.925834 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.925895 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.925916 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.927632 4936 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="18b24896f31b8f587c0f19067d2d9c0ab097baf7bcd9df4f2c87cd5bf009d370" exitCode=0 Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.927691 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"18b24896f31b8f587c0f19067d2d9c0ab097baf7bcd9df4f2c87cd5bf009d370"} Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.927767 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.927812 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.927853 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.927898 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.927879 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.929289 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.929333 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.929351 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.929347 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.929584 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.929603 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.929840 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.929870 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.929892 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.930412 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.930452 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:53 crc kubenswrapper[4936]: I0320 16:00:53.930472 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.377373 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.937695 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"735f7862bf4f4cb8b348eaada2514f5181b123aca2aefcdbfea4baf3f9741bab"} Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.937740 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.937752 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6bdb34b9a91fae7bdc3e0de31eb24c15973fd7f9ddc01b3e0e338183f18a460a"} Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.937770 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"92417113840b0bdc24dbab6d1b660e25eb51497cc45b6563d4c910687419992d"} Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.937833 4936 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.937854 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.937889 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.937909 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.939505 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.939557 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.939569 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.940257 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.940285 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.940299 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.940354 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.940402 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:54 crc kubenswrapper[4936]: I0320 16:00:54.940426 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.772897 4936 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.946431 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8390b843efd1d85122a70952512ed6c15571ab9c94d335fd69721323a095565f"} Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.946496 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"de1fedcde531a606ac0c76657ab81b7fcf8172ca12b1b86b36072fb83c639160"} Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.946575 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.946633 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.948007 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.948045 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.948057 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.948386 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.948411 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:55 crc kubenswrapper[4936]: I0320 16:00:55.948421 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.226988 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.228864 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.228944 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.228967 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.229043 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.483620 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.838423 4936 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.838520 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.884341 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.949070 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.949233 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.950169 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.950212 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.950227 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.951450 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.951613 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:56 crc kubenswrapper[4936]: I0320 16:00:56.951642 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.292396 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.292799 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.295348 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.295450 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.295517 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.496320 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.952018 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.953718 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.953769 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:57 crc kubenswrapper[4936]: I0320 16:00:57.953786 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:58 crc kubenswrapper[4936]: I0320 16:00:58.956487 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:00:58 crc kubenswrapper[4936]: I0320 16:00:58.958729 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:00:58 crc kubenswrapper[4936]: I0320 16:00:58.958784 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:00:58 crc kubenswrapper[4936]: I0320 16:00:58.958797 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:00:59 crc kubenswrapper[4936]: E0320 16:00:59.916034 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.501887 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.502205 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.504007 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.504105 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.504133 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.510091 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.967614 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.969080 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.969192 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:01 crc kubenswrapper[4936]: I0320 16:01:01.969225 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:03 crc kubenswrapper[4936]: I0320 16:01:03.774103 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 20 16:01:03 crc kubenswrapper[4936]: I0320 16:01:03.975107 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 20 16:01:03 crc kubenswrapper[4936]: I0320 16:01:03.977507 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7ad9886cd7aa3c83e1c258952be05467e5d16f14b8701041a35cd852c4c07489" exitCode=255 Mar 20 16:01:03 crc kubenswrapper[4936]: I0320 16:01:03.977627 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7ad9886cd7aa3c83e1c258952be05467e5d16f14b8701041a35cd852c4c07489"} Mar 20 16:01:03 crc kubenswrapper[4936]: I0320 16:01:03.977898 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:03 crc kubenswrapper[4936]: I0320 16:01:03.979278 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:03 crc kubenswrapper[4936]: I0320 16:01:03.979395 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:03 crc kubenswrapper[4936]: I0320 16:01:03.979491 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:03 crc kubenswrapper[4936]: I0320 16:01:03.980245 4936 scope.go:117] "RemoveContainer" containerID="7ad9886cd7aa3c83e1c258952be05467e5d16f14b8701041a35cd852c4c07489" Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.377633 4936 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.377733 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 20 16:01:04 crc kubenswrapper[4936]: E0320 16:01:04.552286 4936 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:04 crc kubenswrapper[4936]: E0320 16:01:04.555784 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.557363 4936 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.557459 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 20 16:01:04 crc kubenswrapper[4936]: W0320 16:01:04.558729 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z Mar 20 16:01:04 crc kubenswrapper[4936]: E0320 16:01:04.558830 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:04 crc kubenswrapper[4936]: E0320 16:01:04.560976 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 16:01:04 crc kubenswrapper[4936]: W0320 16:01:04.562490 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z Mar 20 16:01:04 crc kubenswrapper[4936]: E0320 16:01:04.562565 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:04 crc kubenswrapper[4936]: W0320 16:01:04.564896 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z Mar 20 16:01:04 crc kubenswrapper[4936]: E0320 16:01:04.565017 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:04 crc kubenswrapper[4936]: E0320 16:01:04.565948 4936 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e9804d60b5fe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,LastTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:04 crc kubenswrapper[4936]: W0320 16:01:04.566712 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z Mar 20 16:01:04 crc kubenswrapper[4936]: E0320 16:01:04.566788 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.775290 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:04Z is after 2026-02-23T05:33:13Z Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.981719 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.983592 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218"} Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.983857 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.985239 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.985272 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:04 crc kubenswrapper[4936]: I0320 16:01:04.985283 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:05 crc kubenswrapper[4936]: I0320 16:01:05.776220 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:05Z is after 2026-02-23T05:33:13Z Mar 20 16:01:05 crc kubenswrapper[4936]: I0320 16:01:05.994018 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 16:01:05 crc kubenswrapper[4936]: I0320 16:01:05.995065 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 20 16:01:05 crc kubenswrapper[4936]: I0320 16:01:05.998781 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218" exitCode=255 Mar 20 16:01:05 crc kubenswrapper[4936]: I0320 16:01:05.998880 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218"} Mar 20 16:01:05 crc kubenswrapper[4936]: I0320 16:01:05.998990 4936 scope.go:117] "RemoveContainer" containerID="7ad9886cd7aa3c83e1c258952be05467e5d16f14b8701041a35cd852c4c07489" Mar 20 16:01:05 crc kubenswrapper[4936]: I0320 16:01:05.999336 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:06 crc kubenswrapper[4936]: I0320 16:01:06.001356 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:06 crc kubenswrapper[4936]: I0320 16:01:06.001399 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:06 crc kubenswrapper[4936]: I0320 16:01:06.001414 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:06 crc kubenswrapper[4936]: I0320 16:01:06.002243 4936 scope.go:117] "RemoveContainer" containerID="e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218" Mar 20 16:01:06 crc kubenswrapper[4936]: E0320 16:01:06.002691 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:01:06 crc kubenswrapper[4936]: I0320 16:01:06.620816 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:01:06 crc kubenswrapper[4936]: I0320 16:01:06.778104 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:06Z is after 2026-02-23T05:33:13Z Mar 20 16:01:06 crc kubenswrapper[4936]: I0320 16:01:06.839380 4936 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:01:06 crc kubenswrapper[4936]: I0320 16:01:06.839665 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.003669 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.006723 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.007654 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.007716 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.007734 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.008589 4936 scope.go:117] "RemoveContainer" containerID="e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218" Mar 20 16:01:07 crc kubenswrapper[4936]: E0320 16:01:07.008911 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.034348 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.034631 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.036286 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.036367 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.036388 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.066499 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 20 16:01:07 crc kubenswrapper[4936]: I0320 16:01:07.776615 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:07Z is after 2026-02-23T05:33:13Z Mar 20 16:01:08 crc kubenswrapper[4936]: I0320 16:01:08.010085 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:08 crc kubenswrapper[4936]: I0320 16:01:08.011787 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:08 crc kubenswrapper[4936]: I0320 16:01:08.011866 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:08 crc kubenswrapper[4936]: I0320 16:01:08.011892 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:08 crc kubenswrapper[4936]: I0320 16:01:08.777943 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:08Z is after 2026-02-23T05:33:13Z Mar 20 16:01:09 crc kubenswrapper[4936]: I0320 16:01:09.387367 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:01:09 crc kubenswrapper[4936]: I0320 16:01:09.388508 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:09 crc kubenswrapper[4936]: I0320 16:01:09.390128 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:09 crc kubenswrapper[4936]: I0320 16:01:09.390348 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:09 crc kubenswrapper[4936]: I0320 16:01:09.390497 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:09 crc kubenswrapper[4936]: I0320 16:01:09.391500 4936 scope.go:117] "RemoveContainer" containerID="e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218" Mar 20 16:01:09 crc kubenswrapper[4936]: E0320 16:01:09.392049 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:01:09 crc kubenswrapper[4936]: I0320 16:01:09.394852 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:01:09 crc kubenswrapper[4936]: I0320 16:01:09.776858 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:09Z is after 2026-02-23T05:33:13Z Mar 20 16:01:09 crc kubenswrapper[4936]: E0320 16:01:09.916193 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.016191 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.017840 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.017900 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.017922 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.018772 4936 scope.go:117] "RemoveContainer" containerID="e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218" Mar 20 16:01:10 crc kubenswrapper[4936]: E0320 16:01:10.019047 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.777247 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:10Z is after 2026-02-23T05:33:13Z Mar 20 16:01:10 crc kubenswrapper[4936]: E0320 16:01:10.961988 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:10Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.962269 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.964232 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.964301 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.964327 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:10 crc kubenswrapper[4936]: I0320 16:01:10.964370 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:01:10 crc kubenswrapper[4936]: E0320 16:01:10.969100 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:10Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 16:01:11 crc kubenswrapper[4936]: I0320 16:01:11.777118 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:11Z is after 2026-02-23T05:33:13Z Mar 20 16:01:11 crc kubenswrapper[4936]: W0320 16:01:11.834600 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:11Z is after 2026-02-23T05:33:13Z Mar 20 16:01:11 crc kubenswrapper[4936]: E0320 16:01:11.834722 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:11Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:12 crc kubenswrapper[4936]: I0320 16:01:12.555988 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:01:12 crc kubenswrapper[4936]: I0320 16:01:12.556218 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:12 crc kubenswrapper[4936]: I0320 16:01:12.557890 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:12 crc kubenswrapper[4936]: I0320 16:01:12.557980 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:12 crc kubenswrapper[4936]: I0320 16:01:12.558004 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:12 crc kubenswrapper[4936]: I0320 16:01:12.559050 4936 scope.go:117] "RemoveContainer" containerID="e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218" Mar 20 16:01:12 crc kubenswrapper[4936]: E0320 16:01:12.559406 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:01:12 crc kubenswrapper[4936]: I0320 16:01:12.777710 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:12Z is after 2026-02-23T05:33:13Z Mar 20 16:01:12 crc kubenswrapper[4936]: I0320 16:01:12.872024 4936 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 16:01:12 crc kubenswrapper[4936]: E0320 16:01:12.878579 4936 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:12Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:13 crc kubenswrapper[4936]: I0320 16:01:13.777051 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:13Z is after 2026-02-23T05:33:13Z Mar 20 16:01:13 crc kubenswrapper[4936]: W0320 16:01:13.903512 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:13Z is after 2026-02-23T05:33:13Z Mar 20 16:01:13 crc kubenswrapper[4936]: E0320 16:01:13.903705 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:13Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:14 crc kubenswrapper[4936]: E0320 16:01:14.572613 4936 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:14Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e9804d60b5fe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,LastTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:14 crc kubenswrapper[4936]: I0320 16:01:14.777337 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:14Z is after 2026-02-23T05:33:13Z Mar 20 16:01:15 crc kubenswrapper[4936]: W0320 16:01:15.523728 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:15Z is after 2026-02-23T05:33:13Z Mar 20 16:01:15 crc kubenswrapper[4936]: W0320 16:01:15.523981 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:15Z is after 2026-02-23T05:33:13Z Mar 20 16:01:15 crc kubenswrapper[4936]: E0320 16:01:15.524439 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:15 crc kubenswrapper[4936]: E0320 16:01:15.524293 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:15 crc kubenswrapper[4936]: I0320 16:01:15.777517 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:15Z is after 2026-02-23T05:33:13Z Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.777765 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:16Z is after 2026-02-23T05:33:13Z Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.839633 4936 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.839769 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.839865 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.840113 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.842279 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.842373 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.842396 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.843258 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"161e8a105db0ed826e7d924f9a1556199b4812d8f0c5190581f2286b23c78aa7"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 20 16:01:16 crc kubenswrapper[4936]: I0320 16:01:16.843521 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://161e8a105db0ed826e7d924f9a1556199b4812d8f0c5190581f2286b23c78aa7" gracePeriod=30 Mar 20 16:01:17 crc kubenswrapper[4936]: I0320 16:01:17.042676 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 16:01:17 crc kubenswrapper[4936]: I0320 16:01:17.043292 4936 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="161e8a105db0ed826e7d924f9a1556199b4812d8f0c5190581f2286b23c78aa7" exitCode=255 Mar 20 16:01:17 crc kubenswrapper[4936]: I0320 16:01:17.043366 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"161e8a105db0ed826e7d924f9a1556199b4812d8f0c5190581f2286b23c78aa7"} Mar 20 16:01:17 crc kubenswrapper[4936]: I0320 16:01:17.777594 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:17Z is after 2026-02-23T05:33:13Z Mar 20 16:01:17 crc kubenswrapper[4936]: E0320 16:01:17.969121 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:17Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 20 16:01:17 crc kubenswrapper[4936]: I0320 16:01:17.970193 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:17 crc kubenswrapper[4936]: I0320 16:01:17.972245 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:17 crc kubenswrapper[4936]: I0320 16:01:17.972307 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:17 crc kubenswrapper[4936]: I0320 16:01:17.972352 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:17 crc kubenswrapper[4936]: I0320 16:01:17.972398 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:01:17 crc kubenswrapper[4936]: E0320 16:01:17.978144 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:17Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 16:01:18 crc kubenswrapper[4936]: I0320 16:01:18.052825 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 16:01:18 crc kubenswrapper[4936]: I0320 16:01:18.053714 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"73781019f39b446b4364df7544f87394552b361771d40ff79004107c5856d430"} Mar 20 16:01:18 crc kubenswrapper[4936]: I0320 16:01:18.053952 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:18 crc kubenswrapper[4936]: I0320 16:01:18.055526 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:18 crc kubenswrapper[4936]: I0320 16:01:18.055610 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:18 crc kubenswrapper[4936]: I0320 16:01:18.055629 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:18 crc kubenswrapper[4936]: I0320 16:01:18.775046 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:18Z is after 2026-02-23T05:33:13Z Mar 20 16:01:19 crc kubenswrapper[4936]: I0320 16:01:19.056930 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:19 crc kubenswrapper[4936]: I0320 16:01:19.058597 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:19 crc kubenswrapper[4936]: I0320 16:01:19.058655 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:19 crc kubenswrapper[4936]: I0320 16:01:19.058673 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:19 crc kubenswrapper[4936]: I0320 16:01:19.779472 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:19Z is after 2026-02-23T05:33:13Z Mar 20 16:01:19 crc kubenswrapper[4936]: E0320 16:01:19.916415 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:01:20 crc kubenswrapper[4936]: I0320 16:01:20.779189 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:20Z is after 2026-02-23T05:33:13Z Mar 20 16:01:21 crc kubenswrapper[4936]: I0320 16:01:21.502061 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:01:21 crc kubenswrapper[4936]: I0320 16:01:21.502340 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:21 crc kubenswrapper[4936]: I0320 16:01:21.503819 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:21 crc kubenswrapper[4936]: I0320 16:01:21.503879 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:21 crc kubenswrapper[4936]: I0320 16:01:21.503896 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:21 crc kubenswrapper[4936]: I0320 16:01:21.777296 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:21Z is after 2026-02-23T05:33:13Z Mar 20 16:01:22 crc kubenswrapper[4936]: I0320 16:01:22.775357 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:22Z is after 2026-02-23T05:33:13Z Mar 20 16:01:23 crc kubenswrapper[4936]: I0320 16:01:23.776891 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:23Z is after 2026-02-23T05:33:13Z Mar 20 16:01:23 crc kubenswrapper[4936]: I0320 16:01:23.838282 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:01:23 crc kubenswrapper[4936]: I0320 16:01:23.838499 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:23 crc kubenswrapper[4936]: I0320 16:01:23.840616 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:23 crc kubenswrapper[4936]: I0320 16:01:23.840699 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:23 crc kubenswrapper[4936]: I0320 16:01:23.840719 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:24 crc kubenswrapper[4936]: E0320 16:01:24.578863 4936 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:24Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e9804d60b5fe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,LastTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:24 crc kubenswrapper[4936]: I0320 16:01:24.777320 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:24Z is after 2026-02-23T05:33:13Z Mar 20 16:01:24 crc kubenswrapper[4936]: E0320 16:01:24.975092 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:24Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 20 16:01:24 crc kubenswrapper[4936]: I0320 16:01:24.978267 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:24 crc kubenswrapper[4936]: I0320 16:01:24.980424 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:24 crc kubenswrapper[4936]: I0320 16:01:24.980490 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:24 crc kubenswrapper[4936]: I0320 16:01:24.980514 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:24 crc kubenswrapper[4936]: I0320 16:01:24.980589 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:01:24 crc kubenswrapper[4936]: E0320 16:01:24.985849 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:24Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 16:01:25 crc kubenswrapper[4936]: I0320 16:01:25.777417 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:25Z is after 2026-02-23T05:33:13Z Mar 20 16:01:26 crc kubenswrapper[4936]: I0320 16:01:26.776391 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:26Z is after 2026-02-23T05:33:13Z Mar 20 16:01:26 crc kubenswrapper[4936]: I0320 16:01:26.839311 4936 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:01:26 crc kubenswrapper[4936]: I0320 16:01:26.839413 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 16:01:27 crc kubenswrapper[4936]: I0320 16:01:27.777471 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:27Z is after 2026-02-23T05:33:13Z Mar 20 16:01:27 crc kubenswrapper[4936]: I0320 16:01:27.853564 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:27 crc kubenswrapper[4936]: I0320 16:01:27.855107 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:27 crc kubenswrapper[4936]: I0320 16:01:27.855145 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:27 crc kubenswrapper[4936]: I0320 16:01:27.855155 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:27 crc kubenswrapper[4936]: I0320 16:01:27.855978 4936 scope.go:117] "RemoveContainer" containerID="e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218" Mar 20 16:01:28 crc kubenswrapper[4936]: I0320 16:01:28.777476 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:28Z is after 2026-02-23T05:33:13Z Mar 20 16:01:28 crc kubenswrapper[4936]: I0320 16:01:28.948008 4936 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 16:01:28 crc kubenswrapper[4936]: E0320 16:01:28.954294 4936 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:28Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:28 crc kubenswrapper[4936]: E0320 16:01:28.955907 4936 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.090598 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.091493 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.094259 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="225f64e91792b1df1791fde316604ab759c7bac5b8ee66bc7fe70aa56eec00f0" exitCode=255 Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.094322 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"225f64e91792b1df1791fde316604ab759c7bac5b8ee66bc7fe70aa56eec00f0"} Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.094376 4936 scope.go:117] "RemoveContainer" containerID="e7caeab39ad7a3f8f8674c5c470b9662ddaf96172e60f3a80c943ed4a720d218" Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.094701 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.096029 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.096096 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.096116 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.098049 4936 scope.go:117] "RemoveContainer" containerID="225f64e91792b1df1791fde316604ab759c7bac5b8ee66bc7fe70aa56eec00f0" Mar 20 16:01:29 crc kubenswrapper[4936]: E0320 16:01:29.098567 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:01:29 crc kubenswrapper[4936]: I0320 16:01:29.775210 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:29Z is after 2026-02-23T05:33:13Z Mar 20 16:01:29 crc kubenswrapper[4936]: E0320 16:01:29.916899 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:01:30 crc kubenswrapper[4936]: I0320 16:01:30.100686 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 16:01:30 crc kubenswrapper[4936]: I0320 16:01:30.776137 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:30Z is after 2026-02-23T05:33:13Z Mar 20 16:01:31 crc kubenswrapper[4936]: W0320 16:01:31.017370 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:31Z is after 2026-02-23T05:33:13Z Mar 20 16:01:31 crc kubenswrapper[4936]: E0320 16:01:31.017528 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:31Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:31 crc kubenswrapper[4936]: I0320 16:01:31.776787 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:31Z is after 2026-02-23T05:33:13Z Mar 20 16:01:31 crc kubenswrapper[4936]: E0320 16:01:31.985013 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:31Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 20 16:01:31 crc kubenswrapper[4936]: I0320 16:01:31.986482 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:31 crc kubenswrapper[4936]: I0320 16:01:31.991078 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:31 crc kubenswrapper[4936]: I0320 16:01:31.991196 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:31 crc kubenswrapper[4936]: I0320 16:01:31.991269 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:31 crc kubenswrapper[4936]: I0320 16:01:31.991377 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:01:31 crc kubenswrapper[4936]: E0320 16:01:31.994115 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:31Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 16:01:32 crc kubenswrapper[4936]: I0320 16:01:32.555763 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:01:32 crc kubenswrapper[4936]: I0320 16:01:32.556319 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:32 crc kubenswrapper[4936]: I0320 16:01:32.558292 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:32 crc kubenswrapper[4936]: I0320 16:01:32.558368 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:32 crc kubenswrapper[4936]: I0320 16:01:32.558389 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:32 crc kubenswrapper[4936]: I0320 16:01:32.559451 4936 scope.go:117] "RemoveContainer" containerID="225f64e91792b1df1791fde316604ab759c7bac5b8ee66bc7fe70aa56eec00f0" Mar 20 16:01:32 crc kubenswrapper[4936]: E0320 16:01:32.559838 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:01:32 crc kubenswrapper[4936]: I0320 16:01:32.777251 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:32Z is after 2026-02-23T05:33:13Z Mar 20 16:01:33 crc kubenswrapper[4936]: I0320 16:01:33.777143 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:33Z is after 2026-02-23T05:33:13Z Mar 20 16:01:34 crc kubenswrapper[4936]: W0320 16:01:34.457867 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:34Z is after 2026-02-23T05:33:13Z Mar 20 16:01:34 crc kubenswrapper[4936]: E0320 16:01:34.458032 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:34Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:34 crc kubenswrapper[4936]: E0320 16:01:34.585416 4936 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:34Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e9804d60b5fe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,LastTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:34 crc kubenswrapper[4936]: I0320 16:01:34.775248 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:34Z is after 2026-02-23T05:33:13Z Mar 20 16:01:35 crc kubenswrapper[4936]: I0320 16:01:35.776432 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:35Z is after 2026-02-23T05:33:13Z Mar 20 16:01:36 crc kubenswrapper[4936]: I0320 16:01:36.619903 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:01:36 crc kubenswrapper[4936]: I0320 16:01:36.620101 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:36 crc kubenswrapper[4936]: I0320 16:01:36.626833 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:36 crc kubenswrapper[4936]: I0320 16:01:36.626898 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:36 crc kubenswrapper[4936]: I0320 16:01:36.626917 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:36 crc kubenswrapper[4936]: I0320 16:01:36.627780 4936 scope.go:117] "RemoveContainer" containerID="225f64e91792b1df1791fde316604ab759c7bac5b8ee66bc7fe70aa56eec00f0" Mar 20 16:01:36 crc kubenswrapper[4936]: E0320 16:01:36.628090 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:01:36 crc kubenswrapper[4936]: I0320 16:01:36.776637 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:36Z is after 2026-02-23T05:33:13Z Mar 20 16:01:36 crc kubenswrapper[4936]: I0320 16:01:36.838719 4936 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:01:36 crc kubenswrapper[4936]: I0320 16:01:36.838815 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 16:01:36 crc kubenswrapper[4936]: W0320 16:01:36.961227 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:36Z is after 2026-02-23T05:33:13Z Mar 20 16:01:36 crc kubenswrapper[4936]: E0320 16:01:36.961364 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:36Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:36 crc kubenswrapper[4936]: W0320 16:01:36.961720 4936 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:36Z is after 2026-02-23T05:33:13Z Mar 20 16:01:36 crc kubenswrapper[4936]: E0320 16:01:36.961860 4936 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:36Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 16:01:37 crc kubenswrapper[4936]: I0320 16:01:37.777370 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:37Z is after 2026-02-23T05:33:13Z Mar 20 16:01:38 crc kubenswrapper[4936]: I0320 16:01:38.777116 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:38Z is after 2026-02-23T05:33:13Z Mar 20 16:01:38 crc kubenswrapper[4936]: E0320 16:01:38.991227 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:38Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 20 16:01:38 crc kubenswrapper[4936]: I0320 16:01:38.994415 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:38 crc kubenswrapper[4936]: I0320 16:01:38.996321 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:38 crc kubenswrapper[4936]: I0320 16:01:38.996383 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:38 crc kubenswrapper[4936]: I0320 16:01:38.996397 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:38 crc kubenswrapper[4936]: I0320 16:01:38.996430 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:01:39 crc kubenswrapper[4936]: E0320 16:01:39.000903 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:39Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 16:01:39 crc kubenswrapper[4936]: I0320 16:01:39.775572 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:39Z is after 2026-02-23T05:33:13Z Mar 20 16:01:39 crc kubenswrapper[4936]: E0320 16:01:39.917141 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:01:40 crc kubenswrapper[4936]: I0320 16:01:40.777216 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:01:40Z is after 2026-02-23T05:33:13Z Mar 20 16:01:40 crc kubenswrapper[4936]: I0320 16:01:40.891047 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 16:01:40 crc kubenswrapper[4936]: I0320 16:01:40.891774 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:40 crc kubenswrapper[4936]: I0320 16:01:40.893647 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:40 crc kubenswrapper[4936]: I0320 16:01:40.893855 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:40 crc kubenswrapper[4936]: I0320 16:01:40.894034 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:41 crc kubenswrapper[4936]: I0320 16:01:41.776630 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:42 crc kubenswrapper[4936]: I0320 16:01:42.776362 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:43 crc kubenswrapper[4936]: I0320 16:01:43.779400 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.594142 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d60b5fe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,LastTimestamp:2026-03-20 16:00:49.768914912 +0000 UTC m=+0.715282767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.601378 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a275a1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829148065 +0000 UTC m=+0.775515890,LastTimestamp:2026-03-20 16:00:49.829148065 +0000 UTC m=+0.775515890,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.608822 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a304a4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829184676 +0000 UTC m=+0.775552501,LastTimestamp:2026-03-20 16:00:49.829184676 +0000 UTC m=+0.775552501,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.616760 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a34073 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829199987 +0000 UTC m=+0.775567812,LastTimestamp:2026-03-20 16:00:49.829199987 +0000 UTC m=+0.775567812,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.622268 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804de641c87 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.908948103 +0000 UTC m=+0.855315918,LastTimestamp:2026-03-20 16:00:49.908948103 +0000 UTC m=+0.855315918,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.629284 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a275a1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a275a1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829148065 +0000 UTC m=+0.775515890,LastTimestamp:2026-03-20 16:00:49.956441979 +0000 UTC m=+0.902809804,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.635410 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a304a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a304a4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829184676 +0000 UTC m=+0.775552501,LastTimestamp:2026-03-20 16:00:49.95647915 +0000 UTC m=+0.902846985,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.642810 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a34073\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a34073 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829199987 +0000 UTC m=+0.775567812,LastTimestamp:2026-03-20 16:00:49.95649133 +0000 UTC m=+0.902859155,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.648604 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a275a1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a275a1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829148065 +0000 UTC m=+0.775515890,LastTimestamp:2026-03-20 16:00:49.957869442 +0000 UTC m=+0.904237307,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.653449 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a304a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a304a4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829184676 +0000 UTC m=+0.775552501,LastTimestamp:2026-03-20 16:00:49.957910503 +0000 UTC m=+0.904278358,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.660661 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a34073\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a34073 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829199987 +0000 UTC m=+0.775567812,LastTimestamp:2026-03-20 16:00:49.957937234 +0000 UTC m=+0.904305089,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.667580 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a275a1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a275a1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829148065 +0000 UTC m=+0.775515890,LastTimestamp:2026-03-20 16:00:49.958617875 +0000 UTC m=+0.904985700,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.673115 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a304a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a304a4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829184676 +0000 UTC m=+0.775552501,LastTimestamp:2026-03-20 16:00:49.958666526 +0000 UTC m=+0.905034351,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.681762 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a34073\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a34073 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829199987 +0000 UTC m=+0.775567812,LastTimestamp:2026-03-20 16:00:49.958679627 +0000 UTC m=+0.905047452,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.688928 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a275a1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a275a1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829148065 +0000 UTC m=+0.775515890,LastTimestamp:2026-03-20 16:00:49.959620925 +0000 UTC m=+0.905988780,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.693457 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a304a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a304a4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829184676 +0000 UTC m=+0.775552501,LastTimestamp:2026-03-20 16:00:49.959649686 +0000 UTC m=+0.906017541,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.700301 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a34073\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a34073 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829199987 +0000 UTC m=+0.775567812,LastTimestamp:2026-03-20 16:00:49.959665817 +0000 UTC m=+0.906033672,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.705020 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a275a1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a275a1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829148065 +0000 UTC m=+0.775515890,LastTimestamp:2026-03-20 16:00:49.960874184 +0000 UTC m=+0.907242039,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.707794 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a304a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a304a4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829184676 +0000 UTC m=+0.775552501,LastTimestamp:2026-03-20 16:00:49.960902074 +0000 UTC m=+0.907269929,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.712107 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a34073\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a34073 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829199987 +0000 UTC m=+0.775567812,LastTimestamp:2026-03-20 16:00:49.960922685 +0000 UTC m=+0.907290540,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.715995 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a275a1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a275a1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829148065 +0000 UTC m=+0.775515890,LastTimestamp:2026-03-20 16:00:49.962266406 +0000 UTC m=+0.908634231,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.722200 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a275a1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a275a1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829148065 +0000 UTC m=+0.775515890,LastTimestamp:2026-03-20 16:00:49.962293947 +0000 UTC m=+0.908661772,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.729111 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a304a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a304a4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829184676 +0000 UTC m=+0.775552501,LastTimestamp:2026-03-20 16:00:49.962301757 +0000 UTC m=+0.908669582,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.735834 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a304a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a304a4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829184676 +0000 UTC m=+0.775552501,LastTimestamp:2026-03-20 16:00:49.962310457 +0000 UTC m=+0.908678282,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.743538 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e9804d9a34073\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e9804d9a34073 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:49.829199987 +0000 UTC m=+0.775567812,LastTimestamp:2026-03-20 16:00:49.962320157 +0000 UTC m=+0.908687982,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.753413 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e9804fa461193 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:50.376741267 +0000 UTC m=+1.323109082,LastTimestamp:2026-03-20 16:00:50.376741267 +0000 UTC m=+1.323109082,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.761283 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e9804fa4720fd openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:50.376810749 +0000 UTC m=+1.323178594,LastTimestamp:2026-03-20 16:00:50.376810749 +0000 UTC m=+1.323178594,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.768887 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9804faebf3f2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:50.387612658 +0000 UTC m=+1.333980513,LastTimestamp:2026-03-20 16:00:50.387612658 +0000 UTC m=+1.333980513,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.776197 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e9804fb502bff openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:50.394180607 +0000 UTC m=+1.340548452,LastTimestamp:2026-03-20 16:00:50.394180607 +0000 UTC m=+1.340548452,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: I0320 16:01:44.776455 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.781696 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e9804fc3926fc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:50.409449212 +0000 UTC m=+1.355817057,LastTimestamp:2026-03-20 16:00:50.409449212 +0000 UTC m=+1.355817057,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.788682 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e98052f24ce81 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.263753857 +0000 UTC m=+2.210121712,LastTimestamp:2026-03-20 16:00:51.263753857 +0000 UTC m=+2.210121712,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.795281 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e98052f416a2a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.265628714 +0000 UTC m=+2.211996529,LastTimestamp:2026-03-20 16:00:51.265628714 +0000 UTC m=+2.211996529,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.800386 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98052f52f097 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.266777239 +0000 UTC m=+2.213145064,LastTimestamp:2026-03-20 16:00:51.266777239 +0000 UTC m=+2.213145064,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.806671 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e98052f8dc611 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.270632977 +0000 UTC m=+2.217000792,LastTimestamp:2026-03-20 16:00:51.270632977 +0000 UTC m=+2.217000792,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.812882 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e98053008307b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.278655611 +0000 UTC m=+2.225023436,LastTimestamp:2026-03-20 16:00:51.278655611 +0000 UTC m=+2.225023436,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.819936 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e9805303f8909 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.282282761 +0000 UTC m=+2.228650586,LastTimestamp:2026-03-20 16:00:51.282282761 +0000 UTC m=+2.228650586,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.824715 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e980530607043 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.284439107 +0000 UTC m=+2.230806932,LastTimestamp:2026-03-20 16:00:51.284439107 +0000 UTC m=+2.230806932,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.830886 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e98053090f309 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.287618313 +0000 UTC m=+2.233986148,LastTimestamp:2026-03-20 16:00:51.287618313 +0000 UTC m=+2.233986148,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.837328 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e9805309b4e5e openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.288297054 +0000 UTC m=+2.234664879,LastTimestamp:2026-03-20 16:00:51.288297054 +0000 UTC m=+2.234664879,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.843612 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e9805309ccfed openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.288395757 +0000 UTC m=+2.234763582,LastTimestamp:2026-03-20 16:00:51.288395757 +0000 UTC m=+2.234763582,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.849878 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e980531686bb3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.301739443 +0000 UTC m=+2.248107278,LastTimestamp:2026-03-20 16:00:51.301739443 +0000 UTC m=+2.248107278,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.857570 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e980544184f06 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.615256326 +0000 UTC m=+2.561624141,LastTimestamp:2026-03-20 16:00:51.615256326 +0000 UTC m=+2.561624141,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.861823 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e98054519f34d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.632141133 +0000 UTC m=+2.578508988,LastTimestamp:2026-03-20 16:00:51.632141133 +0000 UTC m=+2.578508988,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.868492 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e980545381e41 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.634118209 +0000 UTC m=+2.580486064,LastTimestamp:2026-03-20 16:00:51.634118209 +0000 UTC m=+2.580486064,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.876522 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98055436326c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.88565054 +0000 UTC m=+2.832018385,LastTimestamp:2026-03-20 16:00:51.88565054 +0000 UTC m=+2.832018385,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.884418 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805548ee23a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.891462714 +0000 UTC m=+2.837830559,LastTimestamp:2026-03-20 16:00:51.891462714 +0000 UTC m=+2.837830559,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.891419 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e9805550b9adc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.899636444 +0000 UTC m=+2.846004269,LastTimestamp:2026-03-20 16:00:51.899636444 +0000 UTC m=+2.846004269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.898392 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e9805551193cf openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.900027855 +0000 UTC m=+2.846395690,LastTimestamp:2026-03-20 16:00:51.900027855 +0000 UTC m=+2.846395690,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.904803 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e9805551a4039 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.900596281 +0000 UTC m=+2.846964186,LastTimestamp:2026-03-20 16:00:51.900596281 +0000 UTC m=+2.846964186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.910876 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e9805568eaf30 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.92500408 +0000 UTC m=+2.871371925,LastTimestamp:2026-03-20 16:00:51.92500408 +0000 UTC m=+2.871371925,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.914871 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e980556a6120a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.926536714 +0000 UTC m=+2.872904539,LastTimestamp:2026-03-20 16:00:51.926536714 +0000 UTC m=+2.872904539,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.920281 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e980565b4f6dc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.179171036 +0000 UTC m=+3.125538841,LastTimestamp:2026-03-20 16:00:52.179171036 +0000 UTC m=+3.125538841,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.927246 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e980565b73a9a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.17931945 +0000 UTC m=+3.125687265,LastTimestamp:2026-03-20 16:00:52.17931945 +0000 UTC m=+3.125687265,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.934006 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e980565c87532 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.180448562 +0000 UTC m=+3.126816377,LastTimestamp:2026-03-20 16:00:52.180448562 +0000 UTC m=+3.126816377,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.941138 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e980565cb2c5f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.180626527 +0000 UTC m=+3.126994352,LastTimestamp:2026-03-20 16:00:52.180626527 +0000 UTC m=+3.126994352,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.946642 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e980565d3e657 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.181198423 +0000 UTC m=+3.127566238,LastTimestamp:2026-03-20 16:00:52.181198423 +0000 UTC m=+3.127566238,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.954019 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e980566927b2d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.193688365 +0000 UTC m=+3.140056180,LastTimestamp:2026-03-20 16:00:52.193688365 +0000 UTC m=+3.140056180,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.961113 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e980566a7fb1f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.195097375 +0000 UTC m=+3.141465180,LastTimestamp:2026-03-20 16:00:52.195097375 +0000 UTC m=+3.141465180,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.968121 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e980566c96dfb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.197289467 +0000 UTC m=+3.143657282,LastTimestamp:2026-03-20 16:00:52.197289467 +0000 UTC m=+3.143657282,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.974746 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e980566fcecf8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.200664312 +0000 UTC m=+3.147032127,LastTimestamp:2026-03-20 16:00:52.200664312 +0000 UTC m=+3.147032127,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.979458 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98056739efbb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.204662715 +0000 UTC m=+3.151030530,LastTimestamp:2026-03-20 16:00:52.204662715 +0000 UTC m=+3.151030530,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.986624 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e980567ce3874 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.21438066 +0000 UTC m=+3.160748475,LastTimestamp:2026-03-20 16:00:52.21438066 +0000 UTC m=+3.160748475,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.992923 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e9805685d3f64 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.223754084 +0000 UTC m=+3.170121899,LastTimestamp:2026-03-20 16:00:52.223754084 +0000 UTC m=+3.170121899,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:44 crc kubenswrapper[4936]: E0320 16:01:44.999814 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e980571aae5ea openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.37983793 +0000 UTC m=+3.326205745,LastTimestamp:2026-03-20 16:00:52.37983793 +0000 UTC m=+3.326205745,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.006268 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e980571cf1525 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.382209317 +0000 UTC m=+3.328577132,LastTimestamp:2026-03-20 16:00:52.382209317 +0000 UTC m=+3.328577132,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.012392 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e98057273bb16 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.392999702 +0000 UTC m=+3.339367517,LastTimestamp:2026-03-20 16:00:52.392999702 +0000 UTC m=+3.339367517,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.018512 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e9805728b9424 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.394562596 +0000 UTC m=+3.340930421,LastTimestamp:2026-03-20 16:00:52.394562596 +0000 UTC m=+3.340930421,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.025260 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e980572f595a0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.401509792 +0000 UTC m=+3.347877617,LastTimestamp:2026-03-20 16:00:52.401509792 +0000 UTC m=+3.347877617,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.032323 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e9805730acee7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.402900711 +0000 UTC m=+3.349268536,LastTimestamp:2026-03-20 16:00:52.402900711 +0000 UTC m=+3.349268536,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.038274 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e9805807d60f2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.62851301 +0000 UTC m=+3.574880825,LastTimestamp:2026-03-20 16:00:52.62851301 +0000 UTC m=+3.574880825,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.044481 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e980580a8de59 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.631363161 +0000 UTC m=+3.577730976,LastTimestamp:2026-03-20 16:00:52.631363161 +0000 UTC m=+3.577730976,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.052400 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e9805815597e2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.64268285 +0000 UTC m=+3.589050665,LastTimestamp:2026-03-20 16:00:52.64268285 +0000 UTC m=+3.589050665,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.058457 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e980581687227 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.643918375 +0000 UTC m=+3.590286200,LastTimestamp:2026-03-20 16:00:52.643918375 +0000 UTC m=+3.590286200,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.064655 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e98058185602e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.645814318 +0000 UTC m=+3.592182123,LastTimestamp:2026-03-20 16:00:52.645814318 +0000 UTC m=+3.592182123,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.070840 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98058cc7aacb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.834708171 +0000 UTC m=+3.781075986,LastTimestamp:2026-03-20 16:00:52.834708171 +0000 UTC m=+3.781075986,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.077918 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98058d9a6475 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.848518261 +0000 UTC m=+3.794886066,LastTimestamp:2026-03-20 16:00:52.848518261 +0000 UTC m=+3.794886066,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.082906 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98058db13cd3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.850015443 +0000 UTC m=+3.796383258,LastTimestamp:2026-03-20 16:00:52.850015443 +0000 UTC m=+3.796383258,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.089304 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e980590af47e2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.90021885 +0000 UTC m=+3.846586655,LastTimestamp:2026-03-20 16:00:52.90021885 +0000 UTC m=+3.846586655,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.097770 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98059ceb7c15 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:53.105490965 +0000 UTC m=+4.051858780,LastTimestamp:2026-03-20 16:00:53.105490965 +0000 UTC m=+4.051858780,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.105322 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98059daea132 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:53.118279986 +0000 UTC m=+4.064647801,LastTimestamp:2026-03-20 16:00:53.118279986 +0000 UTC m=+4.064647801,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.112346 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e98059fd62fde openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:53.154426846 +0000 UTC m=+4.100794661,LastTimestamp:2026-03-20 16:00:53.154426846 +0000 UTC m=+4.100794661,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.119125 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805a0719008 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:53.164609544 +0000 UTC m=+4.110977359,LastTimestamp:2026-03-20 16:00:53.164609544 +0000 UTC m=+4.110977359,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.122102 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805ce33db05 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:53.932317445 +0000 UTC m=+4.878685300,LastTimestamp:2026-03-20 16:00:53.932317445 +0000 UTC m=+4.878685300,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.128688 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805dca8e006 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.174867462 +0000 UTC m=+5.121235327,LastTimestamp:2026-03-20 16:00:54.174867462 +0000 UTC m=+5.121235327,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.134923 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805dd743c8e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.188194958 +0000 UTC m=+5.134562803,LastTimestamp:2026-03-20 16:00:54.188194958 +0000 UTC m=+5.134562803,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.141579 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805dd902f9a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.19002665 +0000 UTC m=+5.136394505,LastTimestamp:2026-03-20 16:00:54.19002665 +0000 UTC m=+5.136394505,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.148867 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805ec4946eb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.437037803 +0000 UTC m=+5.383405618,LastTimestamp:2026-03-20 16:00:54.437037803 +0000 UTC m=+5.383405618,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.157273 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805ed6b2eec openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.4560371 +0000 UTC m=+5.402404915,LastTimestamp:2026-03-20 16:00:54.4560371 +0000 UTC m=+5.402404915,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.164533 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805ed854de8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.457748968 +0000 UTC m=+5.404116773,LastTimestamp:2026-03-20 16:00:54.457748968 +0000 UTC m=+5.404116773,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.170192 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805fad1bb60 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.680861536 +0000 UTC m=+5.627229391,LastTimestamp:2026-03-20 16:00:54.680861536 +0000 UTC m=+5.627229391,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.176178 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805fbbbe9b4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.69620882 +0000 UTC m=+5.642576665,LastTimestamp:2026-03-20 16:00:54.69620882 +0000 UTC m=+5.642576665,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.183707 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e9805fbdb5e46 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.698270278 +0000 UTC m=+5.644638093,LastTimestamp:2026-03-20 16:00:54.698270278 +0000 UTC m=+5.644638093,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.192862 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e98060d6ed8e4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:54.993148132 +0000 UTC m=+5.939515957,LastTimestamp:2026-03-20 16:00:54.993148132 +0000 UTC m=+5.939515957,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.198655 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e98060e3c5351 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:55.006614353 +0000 UTC m=+5.952982178,LastTimestamp:2026-03-20 16:00:55.006614353 +0000 UTC m=+5.952982178,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.203492 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e98060e4df383 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:55.007769475 +0000 UTC m=+5.954137300,LastTimestamp:2026-03-20 16:00:55.007769475 +0000 UTC m=+5.954137300,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.209650 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e98061c945f46 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:55.247265606 +0000 UTC m=+6.193633451,LastTimestamp:2026-03-20 16:00:55.247265606 +0000 UTC m=+6.193633451,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.214247 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e98061dc98c62 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:55.267527778 +0000 UTC m=+6.213895633,LastTimestamp:2026-03-20 16:00:55.267527778 +0000 UTC m=+6.213895633,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.221647 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 16:01:45 crc kubenswrapper[4936]: &Event{ObjectMeta:{kube-controller-manager-crc.189e98067b6c84b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 20 16:01:45 crc kubenswrapper[4936]: body: Mar 20 16:01:45 crc kubenswrapper[4936]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:56.838489267 +0000 UTC m=+7.784857072,LastTimestamp:2026-03-20 16:00:56.838489267 +0000 UTC m=+7.784857072,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 16:01:45 crc kubenswrapper[4936]: > Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.226836 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e98067b6e18ed openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:56.838592749 +0000 UTC m=+7.784960564,LastTimestamp:2026-03-20 16:00:56.838592749 +0000 UTC m=+7.784960564,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.233051 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e98058db13cd3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98058db13cd3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:52.850015443 +0000 UTC m=+3.796383258,LastTimestamp:2026-03-20 16:01:03.982138159 +0000 UTC m=+14.928505964,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.237519 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e98059ceb7c15\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98059ceb7c15 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:53.105490965 +0000 UTC m=+4.051858780,LastTimestamp:2026-03-20 16:01:04.185482839 +0000 UTC m=+15.131850664,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.241482 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e98059daea132\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98059daea132 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:53.118279986 +0000 UTC m=+4.064647801,LastTimestamp:2026-03-20 16:01:04.196687306 +0000 UTC m=+15.143055131,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.245715 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 20 16:01:45 crc kubenswrapper[4936]: &Event{ObjectMeta:{kube-apiserver-crc.189e98083ccbdcba openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:6443/livez": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 20 16:01:45 crc kubenswrapper[4936]: body: Mar 20 16:01:45 crc kubenswrapper[4936]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:04.377707706 +0000 UTC m=+15.324075551,LastTimestamp:2026-03-20 16:01:04.377707706 +0000 UTC m=+15.324075551,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 16:01:45 crc kubenswrapper[4936]: > Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.249659 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e98083cccd3e2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:04.377770978 +0000 UTC m=+15.324138823,LastTimestamp:2026-03-20 16:01:04.377770978 +0000 UTC m=+15.324138823,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.255703 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 20 16:01:45 crc kubenswrapper[4936]: &Event{ObjectMeta:{kube-apiserver-crc.189e9808478246e6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 20 16:01:45 crc kubenswrapper[4936]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 16:01:45 crc kubenswrapper[4936]: Mar 20 16:01:45 crc kubenswrapper[4936]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:04.557434598 +0000 UTC m=+15.503802423,LastTimestamp:2026-03-20 16:01:04.557434598 +0000 UTC m=+15.503802423,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 16:01:45 crc kubenswrapper[4936]: > Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.261658 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e980847855a1c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:04.557636124 +0000 UTC m=+15.504003979,LastTimestamp:2026-03-20 16:01:04.557636124 +0000 UTC m=+15.504003979,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.266086 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 16:01:45 crc kubenswrapper[4936]: &Event{ObjectMeta:{kube-controller-manager-crc.189e9808cf87925b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 16:01:45 crc kubenswrapper[4936]: body: Mar 20 16:01:45 crc kubenswrapper[4936]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:06.839482971 +0000 UTC m=+17.785850816,LastTimestamp:2026-03-20 16:01:06.839482971 +0000 UTC m=+17.785850816,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 16:01:45 crc kubenswrapper[4936]: > Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.272635 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e9808cf8bd0b7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:06.839761079 +0000 UTC m=+17.786128924,LastTimestamp:2026-03-20 16:01:06.839761079 +0000 UTC m=+17.786128924,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.280037 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e9808cf87925b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 16:01:45 crc kubenswrapper[4936]: &Event{ObjectMeta:{kube-controller-manager-crc.189e9808cf87925b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 16:01:45 crc kubenswrapper[4936]: body: Mar 20 16:01:45 crc kubenswrapper[4936]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:06.839482971 +0000 UTC m=+17.785850816,LastTimestamp:2026-03-20 16:01:16.839724824 +0000 UTC m=+27.786092679,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 16:01:45 crc kubenswrapper[4936]: > Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.284612 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e9808cf8bd0b7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e9808cf8bd0b7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:06.839761079 +0000 UTC m=+17.786128924,LastTimestamp:2026-03-20 16:01:16.839817397 +0000 UTC m=+27.786185242,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.289453 4936 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e980b23d0d023 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:16.843503651 +0000 UTC m=+27.789871526,LastTimestamp:2026-03-20 16:01:16.843503651 +0000 UTC m=+27.789871526,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.294203 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e980530607043\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e980530607043 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.284439107 +0000 UTC m=+2.230806932,LastTimestamp:2026-03-20 16:01:16.970026559 +0000 UTC m=+27.916394414,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.300883 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e980544184f06\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e980544184f06 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.615256326 +0000 UTC m=+2.561624141,LastTimestamp:2026-03-20 16:01:17.249173043 +0000 UTC m=+28.195540858,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.304966 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e98054519f34d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e98054519f34d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:00:51.632141133 +0000 UTC m=+2.578508988,LastTimestamp:2026-03-20 16:01:17.259657979 +0000 UTC m=+28.206025794,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.312792 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e9808cf87925b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 16:01:45 crc kubenswrapper[4936]: &Event{ObjectMeta:{kube-controller-manager-crc.189e9808cf87925b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 16:01:45 crc kubenswrapper[4936]: body: Mar 20 16:01:45 crc kubenswrapper[4936]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:06.839482971 +0000 UTC m=+17.785850816,LastTimestamp:2026-03-20 16:01:26.83938269 +0000 UTC m=+37.785750535,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 16:01:45 crc kubenswrapper[4936]: > Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.319017 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e9808cf8bd0b7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e9808cf8bd0b7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:06.839761079 +0000 UTC m=+17.786128924,LastTimestamp:2026-03-20 16:01:26.839463292 +0000 UTC m=+37.785831147,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:01:45 crc kubenswrapper[4936]: E0320 16:01:45.327244 4936 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e9808cf87925b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 16:01:45 crc kubenswrapper[4936]: &Event{ObjectMeta:{kube-controller-manager-crc.189e9808cf87925b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 16:01:45 crc kubenswrapper[4936]: body: Mar 20 16:01:45 crc kubenswrapper[4936]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:01:06.839482971 +0000 UTC m=+17.785850816,LastTimestamp:2026-03-20 16:01:36.838782916 +0000 UTC m=+47.785150731,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 16:01:45 crc kubenswrapper[4936]: > Mar 20 16:01:45 crc kubenswrapper[4936]: I0320 16:01:45.778884 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:46 crc kubenswrapper[4936]: E0320 16:01:46.000494 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.001538 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.002899 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.002940 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.002952 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.002979 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:01:46 crc kubenswrapper[4936]: E0320 16:01:46.010255 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.780097 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.839118 4936 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.839349 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.839467 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.839755 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.841605 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.841685 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.841711 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.842614 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"73781019f39b446b4364df7544f87394552b361771d40ff79004107c5856d430"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 20 16:01:46 crc kubenswrapper[4936]: I0320 16:01:46.842791 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://73781019f39b446b4364df7544f87394552b361771d40ff79004107c5856d430" gracePeriod=30 Mar 20 16:01:47 crc kubenswrapper[4936]: I0320 16:01:47.163288 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 20 16:01:47 crc kubenswrapper[4936]: I0320 16:01:47.165476 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 16:01:47 crc kubenswrapper[4936]: I0320 16:01:47.166126 4936 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="73781019f39b446b4364df7544f87394552b361771d40ff79004107c5856d430" exitCode=255 Mar 20 16:01:47 crc kubenswrapper[4936]: I0320 16:01:47.166168 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"73781019f39b446b4364df7544f87394552b361771d40ff79004107c5856d430"} Mar 20 16:01:47 crc kubenswrapper[4936]: I0320 16:01:47.166212 4936 scope.go:117] "RemoveContainer" containerID="161e8a105db0ed826e7d924f9a1556199b4812d8f0c5190581f2286b23c78aa7" Mar 20 16:01:47 crc kubenswrapper[4936]: I0320 16:01:47.778644 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.173614 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.175970 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9266a01670a96a6d697e78d9e523b5cc65ff5697cae6c0475d7fb3dd953ff81a"} Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.176158 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.177766 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.177833 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.177854 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.773713 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.853441 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.855326 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.855364 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.855373 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:48 crc kubenswrapper[4936]: I0320 16:01:48.855925 4936 scope.go:117] "RemoveContainer" containerID="225f64e91792b1df1791fde316604ab759c7bac5b8ee66bc7fe70aa56eec00f0" Mar 20 16:01:48 crc kubenswrapper[4936]: E0320 16:01:48.856108 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:01:49 crc kubenswrapper[4936]: I0320 16:01:49.179363 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:49 crc kubenswrapper[4936]: I0320 16:01:49.180658 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:49 crc kubenswrapper[4936]: I0320 16:01:49.180717 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:49 crc kubenswrapper[4936]: I0320 16:01:49.180741 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:49 crc kubenswrapper[4936]: I0320 16:01:49.779309 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:49 crc kubenswrapper[4936]: E0320 16:01:49.917460 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:01:50 crc kubenswrapper[4936]: I0320 16:01:50.779081 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:51 crc kubenswrapper[4936]: I0320 16:01:51.502385 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:01:51 crc kubenswrapper[4936]: I0320 16:01:51.502710 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:51 crc kubenswrapper[4936]: I0320 16:01:51.504619 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:51 crc kubenswrapper[4936]: I0320 16:01:51.504698 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:51 crc kubenswrapper[4936]: I0320 16:01:51.504714 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:51 crc kubenswrapper[4936]: I0320 16:01:51.777337 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:52 crc kubenswrapper[4936]: I0320 16:01:52.780073 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:53 crc kubenswrapper[4936]: E0320 16:01:53.005233 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.011346 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.012863 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.012972 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.013036 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.013125 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:01:53 crc kubenswrapper[4936]: E0320 16:01:53.016891 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.778503 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.838398 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.838660 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.840347 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.840425 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.840452 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:53 crc kubenswrapper[4936]: I0320 16:01:53.842855 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:01:54 crc kubenswrapper[4936]: I0320 16:01:54.192984 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:01:54 crc kubenswrapper[4936]: I0320 16:01:54.194050 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:01:54 crc kubenswrapper[4936]: I0320 16:01:54.194096 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:01:54 crc kubenswrapper[4936]: I0320 16:01:54.194106 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:01:54 crc kubenswrapper[4936]: I0320 16:01:54.773336 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:55 crc kubenswrapper[4936]: I0320 16:01:55.776305 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:56 crc kubenswrapper[4936]: I0320 16:01:56.775794 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:57 crc kubenswrapper[4936]: I0320 16:01:57.776015 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:58 crc kubenswrapper[4936]: I0320 16:01:58.777638 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:59 crc kubenswrapper[4936]: I0320 16:01:59.777363 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:01:59 crc kubenswrapper[4936]: E0320 16:01:59.918020 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:02:00 crc kubenswrapper[4936]: E0320 16:02:00.013302 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 16:02:00 crc kubenswrapper[4936]: I0320 16:02:00.017396 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:00 crc kubenswrapper[4936]: I0320 16:02:00.019258 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:00 crc kubenswrapper[4936]: I0320 16:02:00.019508 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:00 crc kubenswrapper[4936]: I0320 16:02:00.019564 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:00 crc kubenswrapper[4936]: I0320 16:02:00.019601 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:02:00 crc kubenswrapper[4936]: E0320 16:02:00.027684 4936 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 16:02:00 crc kubenswrapper[4936]: I0320 16:02:00.780588 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:02:00 crc kubenswrapper[4936]: I0320 16:02:00.958317 4936 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 16:02:00 crc kubenswrapper[4936]: I0320 16:02:00.983245 4936 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 20 16:02:01 crc kubenswrapper[4936]: I0320 16:02:01.508315 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:02:01 crc kubenswrapper[4936]: I0320 16:02:01.508530 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:01 crc kubenswrapper[4936]: I0320 16:02:01.510269 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:01 crc kubenswrapper[4936]: I0320 16:02:01.510328 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:01 crc kubenswrapper[4936]: I0320 16:02:01.510340 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:01 crc kubenswrapper[4936]: I0320 16:02:01.780734 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:02:02 crc kubenswrapper[4936]: I0320 16:02:02.778508 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:02:03 crc kubenswrapper[4936]: I0320 16:02:03.778456 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:02:03 crc kubenswrapper[4936]: I0320 16:02:03.853831 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:03 crc kubenswrapper[4936]: I0320 16:02:03.855391 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:03 crc kubenswrapper[4936]: I0320 16:02:03.855440 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:03 crc kubenswrapper[4936]: I0320 16:02:03.855458 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:03 crc kubenswrapper[4936]: I0320 16:02:03.856406 4936 scope.go:117] "RemoveContainer" containerID="225f64e91792b1df1791fde316604ab759c7bac5b8ee66bc7fe70aa56eec00f0" Mar 20 16:02:04 crc kubenswrapper[4936]: I0320 16:02:04.223761 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 16:02:04 crc kubenswrapper[4936]: I0320 16:02:04.226825 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594"} Mar 20 16:02:04 crc kubenswrapper[4936]: I0320 16:02:04.227487 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:04 crc kubenswrapper[4936]: I0320 16:02:04.228511 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:04 crc kubenswrapper[4936]: I0320 16:02:04.228582 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:04 crc kubenswrapper[4936]: I0320 16:02:04.228596 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:04 crc kubenswrapper[4936]: I0320 16:02:04.781749 4936 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 16:02:05 crc kubenswrapper[4936]: I0320 16:02:05.373999 4936 csr.go:261] certificate signing request csr-4blp2 is approved, waiting to be issued Mar 20 16:02:05 crc kubenswrapper[4936]: I0320 16:02:05.382654 4936 csr.go:257] certificate signing request csr-4blp2 is issued Mar 20 16:02:05 crc kubenswrapper[4936]: I0320 16:02:05.393060 4936 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 20 16:02:05 crc kubenswrapper[4936]: I0320 16:02:05.607829 4936 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.237724 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.238986 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.243087 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594" exitCode=255 Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.243149 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594"} Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.243217 4936 scope.go:117] "RemoveContainer" containerID="225f64e91792b1df1791fde316604ab759c7bac5b8ee66bc7fe70aa56eec00f0" Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.243402 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.245441 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.245487 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.245506 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.246651 4936 scope.go:117] "RemoveContainer" containerID="034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594" Mar 20 16:02:06 crc kubenswrapper[4936]: E0320 16:02:06.246984 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.384285 4936 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-19 01:32:04.852342553 +0000 UTC Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.384366 4936 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 5841h29m58.467982927s for next certificate rotation Mar 20 16:02:06 crc kubenswrapper[4936]: I0320 16:02:06.619959 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.028083 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.029858 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.029906 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.029920 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.030072 4936 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.040667 4936 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.041133 4936 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.041201 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.045408 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.045493 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.045520 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.045587 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.045611 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:07Z","lastTransitionTime":"2026-03-20T16:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.067053 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.078434 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.078641 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.078722 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.078812 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.078897 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:07Z","lastTransitionTime":"2026-03-20T16:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.093639 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.104755 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.104889 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.104955 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.105056 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.105145 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:07Z","lastTransitionTime":"2026-03-20T16:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.120976 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.132908 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.132980 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.132997 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.133018 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.133032 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:07Z","lastTransitionTime":"2026-03-20T16:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.149145 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.149284 4936 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.149318 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.249169 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.249423 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.252186 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.253493 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.253579 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.253606 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:07 crc kubenswrapper[4936]: I0320 16:02:07.254502 4936 scope.go:117] "RemoveContainer" containerID="034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.254821 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.349873 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.450434 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.551092 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.651485 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.751713 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.852018 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:07 crc kubenswrapper[4936]: E0320 16:02:07.952451 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.052981 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.153130 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.253392 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.354600 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.455672 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.557476 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.658979 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.760704 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.861772 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:08 crc kubenswrapper[4936]: E0320 16:02:08.962726 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.063349 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.163788 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.264274 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.365329 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.466487 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.567232 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.668009 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.769076 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: I0320 16:02:09.854182 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:09 crc kubenswrapper[4936]: I0320 16:02:09.856288 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:09 crc kubenswrapper[4936]: I0320 16:02:09.856339 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:09 crc kubenswrapper[4936]: I0320 16:02:09.856351 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.870170 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.918763 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:02:09 crc kubenswrapper[4936]: E0320 16:02:09.970906 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.071856 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.172010 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.272522 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.373402 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.474347 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.574993 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.675840 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.775990 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.876997 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:10 crc kubenswrapper[4936]: E0320 16:02:10.978042 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.078959 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.179968 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.280857 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.381866 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.482666 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.583892 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.684310 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.785396 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.885977 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:11 crc kubenswrapper[4936]: E0320 16:02:11.986669 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.087818 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.188478 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.288685 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.389798 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.490264 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: I0320 16:02:12.556172 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:02:12 crc kubenswrapper[4936]: I0320 16:02:12.556449 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:12 crc kubenswrapper[4936]: I0320 16:02:12.558723 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:12 crc kubenswrapper[4936]: I0320 16:02:12.558801 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:12 crc kubenswrapper[4936]: I0320 16:02:12.558826 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:12 crc kubenswrapper[4936]: I0320 16:02:12.560094 4936 scope.go:117] "RemoveContainer" containerID="034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.560420 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.591259 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.691928 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.792991 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.893415 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:12 crc kubenswrapper[4936]: E0320 16:02:12.993819 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.094322 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.195470 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.295674 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.396179 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.496460 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.596672 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.697407 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.797872 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.898387 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:13 crc kubenswrapper[4936]: E0320 16:02:13.999132 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:14 crc kubenswrapper[4936]: E0320 16:02:14.100003 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:14 crc kubenswrapper[4936]: E0320 16:02:14.200690 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:14 crc kubenswrapper[4936]: E0320 16:02:14.301867 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:14 crc kubenswrapper[4936]: E0320 16:02:14.402190 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:14 crc kubenswrapper[4936]: E0320 16:02:14.503219 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:14 crc kubenswrapper[4936]: E0320 16:02:14.604393 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:14 crc kubenswrapper[4936]: E0320 16:02:14.705590 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:14 crc kubenswrapper[4936]: E0320 16:02:14.806054 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:14 crc kubenswrapper[4936]: E0320 16:02:14.906633 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.007644 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.107784 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.208214 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.308641 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.409750 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.510816 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.611005 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.711882 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.812029 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:15 crc kubenswrapper[4936]: E0320 16:02:15.913240 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.013623 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.114151 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.215084 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.315576 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.416638 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.517763 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.618975 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.719397 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.820450 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:16 crc kubenswrapper[4936]: E0320 16:02:16.921379 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.021809 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.122753 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.154635 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.160430 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.160671 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.160787 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.160905 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.161025 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:17Z","lastTransitionTime":"2026-03-20T16:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.181570 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.191794 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.192059 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.192211 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.192361 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.192498 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:17Z","lastTransitionTime":"2026-03-20T16:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.204538 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.213415 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.213480 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.213491 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.213509 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.213522 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:17Z","lastTransitionTime":"2026-03-20T16:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.230003 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.241913 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.242004 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.242033 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.242069 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:17 crc kubenswrapper[4936]: I0320 16:02:17.242096 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:17Z","lastTransitionTime":"2026-03-20T16:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.256760 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.256928 4936 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.256964 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.357912 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.459154 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.559991 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.660653 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.761424 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.862251 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:17 crc kubenswrapper[4936]: E0320 16:02:17.963842 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.064638 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.165283 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.265921 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.366859 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.468080 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.568629 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.670004 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: I0320 16:02:18.695669 4936 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 20 16:02:18 crc kubenswrapper[4936]: I0320 16:02:18.758261 4936 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.771273 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.872099 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:18 crc kubenswrapper[4936]: E0320 16:02:18.972669 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.073085 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.173695 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.274257 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.375379 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.475781 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.576224 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.677045 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.778040 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.878851 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.919358 4936 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 16:02:19 crc kubenswrapper[4936]: E0320 16:02:19.979167 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.079948 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.181000 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.281687 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.382629 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.483453 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.584120 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.684280 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.784977 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.885777 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:20 crc kubenswrapper[4936]: E0320 16:02:20.986141 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.086812 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.187563 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.288608 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.389271 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.490165 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.591615 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.692432 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.792844 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.893075 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:21 crc kubenswrapper[4936]: E0320 16:02:21.994072 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:22 crc kubenswrapper[4936]: E0320 16:02:22.094757 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:22 crc kubenswrapper[4936]: E0320 16:02:22.195739 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:22 crc kubenswrapper[4936]: E0320 16:02:22.296693 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:22 crc kubenswrapper[4936]: E0320 16:02:22.397070 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:22 crc kubenswrapper[4936]: E0320 16:02:22.497205 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:22 crc kubenswrapper[4936]: E0320 16:02:22.598335 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:22 crc kubenswrapper[4936]: E0320 16:02:22.698999 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:22 crc kubenswrapper[4936]: E0320 16:02:22.800100 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:22 crc kubenswrapper[4936]: E0320 16:02:22.900667 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.001297 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.101786 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.202344 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.303494 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.404471 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.505655 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.606689 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.707301 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.807631 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:23 crc kubenswrapper[4936]: I0320 16:02:23.853438 4936 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 16:02:23 crc kubenswrapper[4936]: I0320 16:02:23.854866 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:23 crc kubenswrapper[4936]: I0320 16:02:23.854905 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:23 crc kubenswrapper[4936]: I0320 16:02:23.854916 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:23 crc kubenswrapper[4936]: I0320 16:02:23.855594 4936 scope.go:117] "RemoveContainer" containerID="034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.855780 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:02:23 crc kubenswrapper[4936]: E0320 16:02:23.908630 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.009772 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.110311 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.211633 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.311978 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.412793 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.513751 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: I0320 16:02:24.606872 4936 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.614244 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.714880 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.815457 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:24 crc kubenswrapper[4936]: E0320 16:02:24.915787 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.016496 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.116606 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.217283 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.317648 4936 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.333459 4936 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.420940 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.421035 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.421060 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.421093 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.421115 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:25Z","lastTransitionTime":"2026-03-20T16:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.524895 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.524955 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.524975 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.525000 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.525018 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:25Z","lastTransitionTime":"2026-03-20T16:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.627750 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.627821 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.627841 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.627868 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.627903 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:25Z","lastTransitionTime":"2026-03-20T16:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.731104 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.731188 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.731204 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.731226 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.731240 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:25Z","lastTransitionTime":"2026-03-20T16:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.803149 4936 apiserver.go:52] "Watching apiserver" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.809722 4936 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.809963 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.810389 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.810536 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.810642 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.810748 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.810766 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.810797 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.810836 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.811057 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.811164 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.815784 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.816262 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.816401 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.816461 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.816502 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.816628 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.816699 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.816905 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.819127 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.834989 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.835034 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.835053 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.835077 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.835095 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:25Z","lastTransitionTime":"2026-03-20T16:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.846664 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.865041 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.876504 4936 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.880095 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.889760 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.898049 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.908785 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.919003 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.932644 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.932786 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.932870 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.932949 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933017 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933062 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933087 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933202 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933230 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933257 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933279 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933299 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933319 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933327 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933310 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933339 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933452 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933501 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933621 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933698 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933772 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933816 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933883 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.933972 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.934157 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.934342 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.934478 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.934512 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.934519 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.934544 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.934805 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935031 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935149 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935185 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935363 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935630 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935653 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935694 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935709 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935733 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935769 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935757 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.935802 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936124 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936213 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936274 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936291 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936363 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936397 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936443 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936479 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936520 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936580 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936666 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936701 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936700 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936738 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936775 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936820 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936853 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936887 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936921 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.936968 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937003 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937043 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937077 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937114 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937148 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937621 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937666 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937698 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937728 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937761 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937792 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937829 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937865 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937902 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937932 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937956 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937980 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938003 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938026 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938050 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938074 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938098 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938125 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938136 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938149 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938172 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938176 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938184 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938199 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938205 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938223 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938218 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:25Z","lastTransitionTime":"2026-03-20T16:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938247 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938272 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938298 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938324 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938348 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938371 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938395 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938422 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938446 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938532 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938598 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938632 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938665 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938693 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938736 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938771 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938833 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938864 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938887 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938911 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938939 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938964 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.938990 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939014 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939038 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939065 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939088 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939113 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939137 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939231 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939310 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939345 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939404 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939439 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939478 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939514 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939575 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939613 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939649 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939683 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939718 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939758 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939795 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939841 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939875 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939920 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939947 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939975 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940003 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940027 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940051 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940083 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940114 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940147 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940173 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940198 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940288 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940323 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940367 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940398 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940433 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940475 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940515 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940588 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940639 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940672 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940710 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940751 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940787 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940820 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940850 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940874 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940898 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940929 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940953 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940979 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941014 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941051 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941091 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941120 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941147 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941171 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941205 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941229 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941253 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941281 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941316 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941346 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941379 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941404 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941429 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941454 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941486 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941519 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941587 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941627 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941669 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941704 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941742 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941777 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941812 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941845 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941880 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941917 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.942147 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.942186 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.942220 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.942257 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.942294 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.942334 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.942369 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.942409 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.942451 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943069 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943196 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943248 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943373 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943466 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943511 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943557 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943725 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943774 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943805 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944209 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944406 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944510 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944693 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944739 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945108 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945290 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945334 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945374 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945594 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945631 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945714 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945896 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945942 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946523 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946568 4936 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946592 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946607 4936 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946624 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946639 4936 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946658 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946675 4936 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946692 4936 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946707 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946726 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946740 4936 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946754 4936 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946769 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946788 4936 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946803 4936 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946818 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946836 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937148 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937204 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.937446 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.939787 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.940969 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.952110 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941599 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941879 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943032 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943633 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943950 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.943616 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944189 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944251 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944306 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944609 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944730 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.944799 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945052 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945101 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945130 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941720 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945647 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.945952 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946426 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946676 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.946936 4936 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946851 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.946971 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.947278 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.947287 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.947425 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.947689 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.947692 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.948485 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.948597 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.941560 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.948613 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.948623 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.948857 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.948931 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.949572 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.949626 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.949647 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.950208 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.950245 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.950266 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.950446 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951013 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951187 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951277 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.955400 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.955513 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.956173 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.956352 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951545 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951619 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951771 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951796 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951832 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951896 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951910 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.952316 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.952483 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.952568 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.952624 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.952719 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.952886 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.953676 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.953998 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.954115 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.954160 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.954750 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:26.454712492 +0000 UTC m=+97.401080307 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.953316 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.954935 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.954821 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.951599 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.956978 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.957151 4936 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.960989 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.961387 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.961422 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.961450 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.961505 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.961654 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.961721 4936 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.962054 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.962425 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.962598 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.962675 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.962721 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.963232 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.963198 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.963320 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.963767 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.964109 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.964134 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.964180 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.965170 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.965243 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.965632 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.965884 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.966280 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.966570 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.966693 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.966713 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.966705 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.967484 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.967495 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.967540 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.967658 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.967669 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.967839 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.967792 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.968299 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.969318 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.969483 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.969581 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.969945 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.969961 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.970230 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.970249 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.971356 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.970391 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.970420 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.971073 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.971325 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.972063 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.972424 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.972575 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.972786 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.973167 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.973206 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.973361 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.973387 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.973517 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.973765 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.973823 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.958673 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.975372 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:26.475336376 +0000 UTC m=+97.421704201 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.975800 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:26.475780348 +0000 UTC m=+97.422148173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.978012 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.978053 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.978070 4936 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.978146 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:26.478123532 +0000 UTC m=+97.424491538 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.988844 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.988845 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.989060 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.989102 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.989124 4936 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:25 crc kubenswrapper[4936]: E0320 16:02:25.989192 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:26.489170215 +0000 UTC m=+97.435538260 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.989365 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.990062 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.990693 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.990702 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.990819 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.990919 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.991074 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.991568 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.992166 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.992284 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.992373 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.992565 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.992702 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.993305 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.993433 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.993703 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.993715 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.994412 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.994688 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.995290 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.994709 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.995547 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.995645 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.995750 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.995764 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.996282 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.996715 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.996763 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.996822 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:25 crc kubenswrapper[4936]: I0320 16:02:25.998776 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.000787 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.001275 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.001575 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.002482 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.003158 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.003319 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.004090 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.004273 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.005256 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.005361 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.007722 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.012095 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.012974 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.024444 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.030010 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.036197 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.038967 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.040490 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.040536 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.040567 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.040585 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.040596 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048093 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048148 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048196 4936 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048207 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048219 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048228 4936 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048236 4936 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048246 4936 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048255 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048267 4936 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048278 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048288 4936 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048297 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048306 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048315 4936 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048325 4936 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048335 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048202 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048344 4936 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048378 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048391 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048402 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048412 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048421 4936 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048430 4936 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048427 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048439 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048505 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048519 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048534 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048574 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048587 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048600 4936 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048612 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048623 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048638 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048649 4936 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048661 4936 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048673 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048687 4936 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048704 4936 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048722 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048738 4936 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048752 4936 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048764 4936 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048778 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048791 4936 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048804 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048818 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048830 4936 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048843 4936 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048858 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048871 4936 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048882 4936 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048894 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048906 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048919 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048932 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048944 4936 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048955 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048967 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048978 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.048990 4936 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049001 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049012 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049023 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049034 4936 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049065 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049076 4936 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049088 4936 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049102 4936 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049113 4936 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049125 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049136 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049147 4936 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049159 4936 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049171 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049182 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049194 4936 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049208 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049220 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049233 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049245 4936 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049256 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049268 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049281 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049294 4936 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049306 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049319 4936 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049331 4936 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049343 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049355 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049367 4936 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049379 4936 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049390 4936 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049402 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049413 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049425 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049437 4936 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049448 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049461 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049473 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049485 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049498 4936 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049509 4936 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049521 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049533 4936 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049896 4936 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049911 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049923 4936 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049935 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049947 4936 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049959 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049972 4936 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049985 4936 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.049998 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050013 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050025 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050036 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050049 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050061 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050073 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050084 4936 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050100 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050111 4936 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050122 4936 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050135 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050148 4936 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050159 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050171 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050183 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050194 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050206 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050219 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050231 4936 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050244 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050256 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050268 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050279 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050292 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050303 4936 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050315 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050327 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050340 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050351 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050364 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050376 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050388 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050400 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050415 4936 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050426 4936 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050438 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050450 4936 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050461 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050473 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050485 4936 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050498 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050511 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050526 4936 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050538 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050572 4936 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050587 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050604 4936 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050616 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050627 4936 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050639 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050651 4936 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050662 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050674 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050685 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050697 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050708 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050718 4936 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050730 4936 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050742 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050754 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050765 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050778 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050789 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.050800 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.128453 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.135925 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.141659 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.143445 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.143632 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.143734 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.143844 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.143945 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:26 crc kubenswrapper[4936]: W0320 16:02:26.160791 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-7659130c8fd836fa9ff093d5a7f2e5fcd3d2c761aaf0cecb3252b29d76afda43 WatchSource:0}: Error finding container 7659130c8fd836fa9ff093d5a7f2e5fcd3d2c761aaf0cecb3252b29d76afda43: Status 404 returned error can't find the container with id 7659130c8fd836fa9ff093d5a7f2e5fcd3d2c761aaf0cecb3252b29d76afda43 Mar 20 16:02:26 crc kubenswrapper[4936]: W0320 16:02:26.164564 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-f76eb3384f8fe70150b8de9a380780ec2339614c50209bc141952315c74064dd WatchSource:0}: Error finding container f76eb3384f8fe70150b8de9a380780ec2339614c50209bc141952315c74064dd: Status 404 returned error can't find the container with id f76eb3384f8fe70150b8de9a380780ec2339614c50209bc141952315c74064dd Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.247703 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.247747 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.247761 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.247783 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.247797 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.310533 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1bacfa517db1b18d58fc2b5e5546057f4cfe62386e134515c54863bbeb61a651"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.311739 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f76eb3384f8fe70150b8de9a380780ec2339614c50209bc141952315c74064dd"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.312592 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7659130c8fd836fa9ff093d5a7f2e5fcd3d2c761aaf0cecb3252b29d76afda43"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.351448 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.351530 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.351595 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.351628 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.351683 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.454072 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.454113 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.454123 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.454139 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.454149 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.554494 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.554619 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.554652 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.554679 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554780 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:27.554754097 +0000 UTC m=+98.501121912 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554777 4936 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554824 4936 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554861 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:27.5548489 +0000 UTC m=+98.501216705 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.554855 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554876 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554894 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:27.554875671 +0000 UTC m=+98.501243486 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554910 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554934 4936 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554934 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554961 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554975 4936 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.554990 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:27.554972473 +0000 UTC m=+98.501340368 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.555029 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:27.555009554 +0000 UTC m=+98.501377459 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.556523 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.556583 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.556597 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.556617 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.556630 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.658845 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.658885 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.658895 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.658911 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.658923 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.762356 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.762402 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.762413 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.762429 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.762442 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.853197 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:26 crc kubenswrapper[4936]: E0320 16:02:26.853431 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.865729 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.865781 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.865796 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.865817 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.865830 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.968597 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.968654 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.968672 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.968701 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:26 crc kubenswrapper[4936]: I0320 16:02:26.968715 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:26Z","lastTransitionTime":"2026-03-20T16:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.071790 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.071856 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.071874 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.071899 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.071921 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.174899 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.174964 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.174983 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.175008 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.175026 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.277839 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.277900 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.277913 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.277934 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.277948 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.317744 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f181defb952a5d819f16c4d15dae8152ed8ef8a28dd069741ceaaa8624b26f2d"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.320180 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0336e2858d3743cdc319d39fe7618b33c18f967c516f5fbe628c24957c8ba63d"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.320255 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"85315042e8881946a84f0f06ce88badcfcdcde9303b3ff4edd6930ca35854867"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.335907 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.351155 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.369651 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.380990 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.381018 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.381027 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.381043 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.381055 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.389987 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f181defb952a5d819f16c4d15dae8152ed8ef8a28dd069741ceaaa8624b26f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T16:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.405814 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.405841 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.405850 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.405860 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.405869 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.406351 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.423713 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.423839 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.428464 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.428582 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.428598 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.428610 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.428618 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.444816 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f181defb952a5d819f16c4d15dae8152ed8ef8a28dd069741ceaaa8624b26f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T16:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.444918 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.448984 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.449020 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.449034 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.449055 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.449071 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.463594 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.467607 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.474359 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.474406 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.474417 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.474436 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.474446 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.481651 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0336e2858d3743cdc319d39fe7618b33c18f967c516f5fbe628c24957c8ba63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T16:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85315042e8881946a84f0f06ce88badcfcdcde9303b3ff4edd6930ca35854867\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T16:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.487859 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.491663 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.493020 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.493047 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.493070 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.493086 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.497433 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.508084 4936 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fc88e84-3312-4795-974e-5b099fa998eb\\\",\\\"systemUUID\\\":\\\"f03975ca-8846-4c96-a4d2-0fc5b8167315\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.508293 4936 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.510505 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.510589 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.510611 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.510635 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.510652 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.517196 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.534635 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:27Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.566094 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.566208 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566284 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:29.566258038 +0000 UTC m=+100.512625853 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566320 4936 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.566328 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.566355 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566378 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:29.566364541 +0000 UTC m=+100.512732356 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.566407 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566561 4936 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566606 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:29.566594258 +0000 UTC m=+100.512962073 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566690 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566704 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566707 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566717 4936 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566876 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:29.566865825 +0000 UTC m=+100.513233640 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.566984 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.567027 4936 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.567137 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:29.567107572 +0000 UTC m=+100.513475427 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.613316 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.613369 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.613383 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.613403 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.613417 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.716534 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.716625 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.716635 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.716655 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.716668 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.819800 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.819866 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.819889 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.819920 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.819943 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.855906 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.856032 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.856223 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:27 crc kubenswrapper[4936]: E0320 16:02:27.856300 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.857933 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.858854 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.860260 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.860988 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.862234 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.862881 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.863542 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.864781 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.865505 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.866726 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.867354 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.868488 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.869099 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.869701 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.870669 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.871222 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.872253 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.872682 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.873277 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.874299 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.874824 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.875923 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.876469 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.877534 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.878058 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.878732 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.879913 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.880441 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.881398 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.882081 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.882965 4936 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.883091 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.884783 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.885676 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.886438 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.888017 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.888645 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.889519 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.890163 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.891182 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.891727 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.892677 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.893262 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.894241 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.894707 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.895885 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.896377 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.897461 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.897984 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.898813 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.899352 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.900361 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.900940 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.901396 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.922290 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.922630 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.922706 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.922805 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:27 crc kubenswrapper[4936]: I0320 16:02:27.922864 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:27Z","lastTransitionTime":"2026-03-20T16:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.025775 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.026187 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.026362 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.026493 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.026617 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.105784 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-7vbr8"] Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.106665 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7vbr8" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.111698 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.112201 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.113798 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.129329 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.129383 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.129396 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.129419 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.129432 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.134254 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f181defb952a5d819f16c4d15dae8152ed8ef8a28dd069741ceaaa8624b26f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T16:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:28Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.151202 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:28Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.167768 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0336e2858d3743cdc319d39fe7618b33c18f967c516f5fbe628c24957c8ba63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T16:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85315042e8881946a84f0f06ce88badcfcdcde9303b3ff4edd6930ca35854867\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T16:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:28Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.181882 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:28Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.195869 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7vbr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7573512a-c722-41b6-a9e8-5993b10ece61\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctsfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T16:02:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7vbr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:28Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.212016 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:28Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.227283 4936 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T16:02:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T16:02:28Z is after 2025-08-24T17:21:41Z" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.232292 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.232508 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.232638 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.232749 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.232829 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.273040 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7573512a-c722-41b6-a9e8-5993b10ece61-hosts-file\") pod \"node-resolver-7vbr8\" (UID: \"7573512a-c722-41b6-a9e8-5993b10ece61\") " pod="openshift-dns/node-resolver-7vbr8" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.273085 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctsfc\" (UniqueName: \"kubernetes.io/projected/7573512a-c722-41b6-a9e8-5993b10ece61-kube-api-access-ctsfc\") pod \"node-resolver-7vbr8\" (UID: \"7573512a-c722-41b6-a9e8-5993b10ece61\") " pod="openshift-dns/node-resolver-7vbr8" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.335839 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.335932 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.335947 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.335969 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.335990 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.374323 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7573512a-c722-41b6-a9e8-5993b10ece61-hosts-file\") pod \"node-resolver-7vbr8\" (UID: \"7573512a-c722-41b6-a9e8-5993b10ece61\") " pod="openshift-dns/node-resolver-7vbr8" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.374373 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctsfc\" (UniqueName: \"kubernetes.io/projected/7573512a-c722-41b6-a9e8-5993b10ece61-kube-api-access-ctsfc\") pod \"node-resolver-7vbr8\" (UID: \"7573512a-c722-41b6-a9e8-5993b10ece61\") " pod="openshift-dns/node-resolver-7vbr8" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.374496 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7573512a-c722-41b6-a9e8-5993b10ece61-hosts-file\") pod \"node-resolver-7vbr8\" (UID: \"7573512a-c722-41b6-a9e8-5993b10ece61\") " pod="openshift-dns/node-resolver-7vbr8" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.404208 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctsfc\" (UniqueName: \"kubernetes.io/projected/7573512a-c722-41b6-a9e8-5993b10ece61-kube-api-access-ctsfc\") pod \"node-resolver-7vbr8\" (UID: \"7573512a-c722-41b6-a9e8-5993b10ece61\") " pod="openshift-dns/node-resolver-7vbr8" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.420943 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7vbr8" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.438887 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.438933 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.438946 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.438968 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.438981 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.518065 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-4cxh6"] Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.518608 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-zjjnf"] Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.518904 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.518944 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-sp67w"] Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.519152 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.520108 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.523400 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.528890 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.529052 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.529178 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.529204 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.529306 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.529356 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.529413 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.529604 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.529747 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.531357 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.532081 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.537378 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jr675"] Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.554641 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.559764 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.559826 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.559846 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.559872 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.559891 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.575595 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.576029 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.576143 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.576342 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.578730 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.579037 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.584066 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.666390 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.666447 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.666461 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.666483 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.666499 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.681947 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-run-k8s-cni-cncf-io\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682214 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-var-lib-cni-bin\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682298 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-run-multus-certs\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682369 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-slash\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682434 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-env-overrides\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682508 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-kubelet\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682598 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-systemd-units\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682669 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-log-socket\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682733 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-bin\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682848 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-socket-dir-parent\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.682942 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-systemd\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683023 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-etc-openvswitch\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683096 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-script-lib\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683174 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683259 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-etc-kubernetes\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683335 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-cnibin\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683411 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-config\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683485 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-os-release\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683576 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-system-cni-dir\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683665 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-os-release\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683762 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-daemon-config\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.683929 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-var-lib-openvswitch\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684037 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovn-node-metrics-cert\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684119 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-cnibin\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684199 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-cni-dir\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684272 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-cni-binary-copy\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684343 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-var-lib-kubelet\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684423 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-openvswitch\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684495 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-node-log\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684583 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-ovn-kubernetes\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684742 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzxnj\" (UniqueName: \"kubernetes.io/projected/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-kube-api-access-xzxnj\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684834 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7l84\" (UniqueName: \"kubernetes.io/projected/dc3fb53f-2e69-4e94-bfa6-762afabe9063-kube-api-access-k7l84\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.684923 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dc3fb53f-2e69-4e94-bfa6-762afabe9063-mcd-auth-proxy-config\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685010 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkc42\" (UniqueName: \"kubernetes.io/projected/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-kube-api-access-lkc42\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685133 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-var-lib-cni-multus\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685220 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-hostroot\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685299 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-netns\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685396 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/dc3fb53f-2e69-4e94-bfa6-762afabe9063-rootfs\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685460 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-system-cni-dir\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685494 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e3d9dfab-9151-432d-ab1b-a78c76959b70-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685525 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-run-netns\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685596 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-ovn\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685630 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dc3fb53f-2e69-4e94-bfa6-762afabe9063-proxy-tls\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685671 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685710 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-conf-dir\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685733 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-netd\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685758 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e3d9dfab-9151-432d-ab1b-a78c76959b70-cni-binary-copy\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.685798 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbp5b\" (UniqueName: \"kubernetes.io/projected/e3d9dfab-9151-432d-ab1b-a78c76959b70-kube-api-access-lbp5b\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.743008 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-dwhls"] Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.743534 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.746007 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.746618 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.746856 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.749045 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.769182 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.769230 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.769258 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.769278 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.769290 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786623 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-kubelet\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786687 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-systemd-units\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786714 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-log-socket\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786745 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-bin\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786761 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-kubelet\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786781 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-socket-dir-parent\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786831 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-systemd-units\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786874 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-bin\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786886 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-log-socket\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786840 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.786899 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-socket-dir-parent\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787025 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-systemd\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787056 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-etc-openvswitch\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787072 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-script-lib\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787101 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-etc-kubernetes\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787120 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-cnibin\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787139 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-config\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787150 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-systemd\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787157 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-os-release\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787200 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-etc-kubernetes\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787231 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-system-cni-dir\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787247 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-cnibin\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787212 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-os-release\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787301 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-system-cni-dir\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787278 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-os-release\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787355 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-os-release\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787451 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovn-node-metrics-cert\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787496 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-cnibin\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787439 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787524 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-daemon-config\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787566 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-cnibin\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787571 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-var-lib-openvswitch\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787611 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-openvswitch\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787626 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-var-lib-openvswitch\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787630 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-node-log\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787656 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-node-log\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787660 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-openvswitch\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787675 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-ovn-kubernetes\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787712 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-cni-dir\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787733 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-cni-binary-copy\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787758 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-ovn-kubernetes\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787760 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-var-lib-kubelet\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787789 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-var-lib-kubelet\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787818 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-cni-dir\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787820 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzxnj\" (UniqueName: \"kubernetes.io/projected/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-kube-api-access-xzxnj\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787931 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7l84\" (UniqueName: \"kubernetes.io/projected/dc3fb53f-2e69-4e94-bfa6-762afabe9063-kube-api-access-k7l84\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.787989 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dc3fb53f-2e69-4e94-bfa6-762afabe9063-mcd-auth-proxy-config\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788017 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkc42\" (UniqueName: \"kubernetes.io/projected/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-kube-api-access-lkc42\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788048 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-var-lib-cni-multus\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788068 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-hostroot\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788087 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-netns\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788113 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/dc3fb53f-2e69-4e94-bfa6-762afabe9063-rootfs\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788150 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-system-cni-dir\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788169 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e3d9dfab-9151-432d-ab1b-a78c76959b70-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788190 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-run-netns\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788207 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-ovn\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788227 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dc3fb53f-2e69-4e94-bfa6-762afabe9063-proxy-tls\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788250 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e3d9dfab-9151-432d-ab1b-a78c76959b70-system-cni-dir\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788255 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-hostroot\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788308 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788351 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-netns\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788351 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-var-lib-cni-multus\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788382 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/dc3fb53f-2e69-4e94-bfa6-762afabe9063-rootfs\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788410 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-ovn\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788412 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-run-netns\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788257 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788454 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e3d9dfab-9151-432d-ab1b-a78c76959b70-cni-binary-copy\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788481 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbp5b\" (UniqueName: \"kubernetes.io/projected/e3d9dfab-9151-432d-ab1b-a78c76959b70-kube-api-access-lbp5b\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788508 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-conf-dir\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788532 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-netd\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788581 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-run-multus-certs\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788604 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-slash\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788607 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-config\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788630 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-env-overrides\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788661 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-run-k8s-cni-cncf-io\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788685 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-var-lib-cni-bin\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788757 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-var-lib-cni-bin\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788789 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-conf-dir\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788448 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-etc-openvswitch\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788818 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-netd\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788843 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-script-lib\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788887 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-slash\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788938 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-run-k8s-cni-cncf-io\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788948 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-host-run-multus-certs\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.788961 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-multus-daemon-config\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.789098 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dc3fb53f-2e69-4e94-bfa6-762afabe9063-mcd-auth-proxy-config\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.789283 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e3d9dfab-9151-432d-ab1b-a78c76959b70-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.789306 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-cni-binary-copy\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.789433 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-env-overrides\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.790601 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e3d9dfab-9151-432d-ab1b-a78c76959b70-cni-binary-copy\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.795308 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dc3fb53f-2e69-4e94-bfa6-762afabe9063-proxy-tls\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.805595 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovn-node-metrics-cert\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.806140 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzxnj\" (UniqueName: \"kubernetes.io/projected/6f8c1cbb-9378-440e-87df-1bc4e70bcaeb-kube-api-access-xzxnj\") pod \"multus-zjjnf\" (UID: \"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb\") " pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.807016 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbp5b\" (UniqueName: \"kubernetes.io/projected/e3d9dfab-9151-432d-ab1b-a78c76959b70-kube-api-access-lbp5b\") pod \"multus-additional-cni-plugins-sp67w\" (UID: \"e3d9dfab-9151-432d-ab1b-a78c76959b70\") " pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.809075 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7l84\" (UniqueName: \"kubernetes.io/projected/dc3fb53f-2e69-4e94-bfa6-762afabe9063-kube-api-access-k7l84\") pod \"machine-config-daemon-4cxh6\" (UID: \"dc3fb53f-2e69-4e94-bfa6-762afabe9063\") " pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.811919 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkc42\" (UniqueName: \"kubernetes.io/projected/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-kube-api-access-lkc42\") pod \"ovnkube-node-jr675\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.854106 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.854154 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:02:28 crc kubenswrapper[4936]: E0320 16:02:28.854372 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.871625 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.871672 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.871686 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.871704 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.871715 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.879193 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zjjnf" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.889492 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/69ac761e-0a13-405b-be5a-60ad535b77a7-host\") pod \"node-ca-dwhls\" (UID: \"69ac761e-0a13-405b-be5a-60ad535b77a7\") " pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.889568 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2szfz\" (UniqueName: \"kubernetes.io/projected/69ac761e-0a13-405b-be5a-60ad535b77a7-kube-api-access-2szfz\") pod \"node-ca-dwhls\" (UID: \"69ac761e-0a13-405b-be5a-60ad535b77a7\") " pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.889747 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/69ac761e-0a13-405b-be5a-60ad535b77a7-serviceca\") pod \"node-ca-dwhls\" (UID: \"69ac761e-0a13-405b-be5a-60ad535b77a7\") " pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.894447 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.905020 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc"] Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.905645 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.910691 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.911443 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sp67w" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.918443 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.933157 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7jnpn"] Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.933784 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:28 crc kubenswrapper[4936]: E0320 16:02:28.933855 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7jnpn" podUID="b65a2c6b-8b39-4742-bd96-6c4cbdc768ac" Mar 20 16:02:28 crc kubenswrapper[4936]: W0320 16:02:28.959269 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3d9dfab_9151_432d_ab1b_a78c76959b70.slice/crio-902f885f562005821b131f8b749f0dfbe3b4b94a467878a1f80b6c634a572073 WatchSource:0}: Error finding container 902f885f562005821b131f8b749f0dfbe3b4b94a467878a1f80b6c634a572073: Status 404 returned error can't find the container with id 902f885f562005821b131f8b749f0dfbe3b4b94a467878a1f80b6c634a572073 Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.974713 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.974748 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.974769 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.974792 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.974810 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:28Z","lastTransitionTime":"2026-03-20T16:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.991264 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/69ac761e-0a13-405b-be5a-60ad535b77a7-serviceca\") pod \"node-ca-dwhls\" (UID: \"69ac761e-0a13-405b-be5a-60ad535b77a7\") " pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.991324 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/69ac761e-0a13-405b-be5a-60ad535b77a7-host\") pod \"node-ca-dwhls\" (UID: \"69ac761e-0a13-405b-be5a-60ad535b77a7\") " pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.991373 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2szfz\" (UniqueName: \"kubernetes.io/projected/69ac761e-0a13-405b-be5a-60ad535b77a7-kube-api-access-2szfz\") pod \"node-ca-dwhls\" (UID: \"69ac761e-0a13-405b-be5a-60ad535b77a7\") " pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.991999 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/69ac761e-0a13-405b-be5a-60ad535b77a7-host\") pod \"node-ca-dwhls\" (UID: \"69ac761e-0a13-405b-be5a-60ad535b77a7\") " pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:28 crc kubenswrapper[4936]: I0320 16:02:28.994304 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/69ac761e-0a13-405b-be5a-60ad535b77a7-serviceca\") pod \"node-ca-dwhls\" (UID: \"69ac761e-0a13-405b-be5a-60ad535b77a7\") " pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.019226 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2szfz\" (UniqueName: \"kubernetes.io/projected/69ac761e-0a13-405b-be5a-60ad535b77a7-kube-api-access-2szfz\") pod \"node-ca-dwhls\" (UID: \"69ac761e-0a13-405b-be5a-60ad535b77a7\") " pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.060864 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dwhls" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.078114 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.078159 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.078196 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.078219 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.078242 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:29Z","lastTransitionTime":"2026-03-20T16:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.092760 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqpv8\" (UniqueName: \"kubernetes.io/projected/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-kube-api-access-mqpv8\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.092804 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vgn8\" (UniqueName: \"kubernetes.io/projected/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-kube-api-access-6vgn8\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.092823 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.092848 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.092863 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.092901 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:29 crc kubenswrapper[4936]: W0320 16:02:29.094684 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69ac761e_0a13_405b_be5a_60ad535b77a7.slice/crio-7c73118fb8cfaf84a9942e5dcf8d4f5b367fe6928b839eb9675f6afa89835958 WatchSource:0}: Error finding container 7c73118fb8cfaf84a9942e5dcf8d4f5b367fe6928b839eb9675f6afa89835958: Status 404 returned error can't find the container with id 7c73118fb8cfaf84a9942e5dcf8d4f5b367fe6928b839eb9675f6afa89835958 Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.182087 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.182141 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.182155 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.182176 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.182190 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:29Z","lastTransitionTime":"2026-03-20T16:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.193854 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.193915 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqpv8\" (UniqueName: \"kubernetes.io/projected/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-kube-api-access-mqpv8\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.193948 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vgn8\" (UniqueName: \"kubernetes.io/projected/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-kube-api-access-6vgn8\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.193972 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.193996 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.194027 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.194035 4936 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.194126 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs podName:b65a2c6b-8b39-4742-bd96-6c4cbdc768ac nodeName:}" failed. No retries permitted until 2026-03-20 16:02:29.694100391 +0000 UTC m=+100.640468206 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs") pod "network-metrics-daemon-7jnpn" (UID: "b65a2c6b-8b39-4742-bd96-6c4cbdc768ac") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.195162 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.195610 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.200181 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.212006 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vgn8\" (UniqueName: \"kubernetes.io/projected/6c4d53e5-257f-435a-88f1-9c2e5ad1f10c-kube-api-access-6vgn8\") pod \"ovnkube-control-plane-749d76644c-kpbcc\" (UID: \"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.228636 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqpv8\" (UniqueName: \"kubernetes.io/projected/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-kube-api-access-mqpv8\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.267278 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.284956 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.285011 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.285025 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.285046 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.285060 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:29Z","lastTransitionTime":"2026-03-20T16:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:29 crc kubenswrapper[4936]: W0320 16:02:29.295925 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c4d53e5_257f_435a_88f1_9c2e5ad1f10c.slice/crio-263a4c5a9e390ca57b86b14e51e0c4dd5f7368dee4a20f210403c25b3c617ff0 WatchSource:0}: Error finding container 263a4c5a9e390ca57b86b14e51e0c4dd5f7368dee4a20f210403c25b3c617ff0: Status 404 returned error can't find the container with id 263a4c5a9e390ca57b86b14e51e0c4dd5f7368dee4a20f210403c25b3c617ff0 Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.329523 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7vbr8" event={"ID":"7573512a-c722-41b6-a9e8-5993b10ece61","Type":"ContainerStarted","Data":"54ef1deef31f1b852d214c5f385dcff8d5acf43a71f3c062c1c9ca0a2c7de9e6"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.329617 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7vbr8" event={"ID":"7573512a-c722-41b6-a9e8-5993b10ece61","Type":"ContainerStarted","Data":"bf78969fb2a81827e152dc01e2322d48477fdce8fbbe889656e059753ffc40f8"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.332265 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerStarted","Data":"49e88b71c957a8ebb056938c31f940d0e00c7629c646b8bfd2d4b1a0ce3b5232"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.332291 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerStarted","Data":"902f885f562005821b131f8b749f0dfbe3b4b94a467878a1f80b6c634a572073"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.335899 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a0e05de5fe6256c8c30e0ba627fd1ba8bffb49e99a6dff910d41bc87477c70ec"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.337246 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" event={"ID":"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c","Type":"ContainerStarted","Data":"263a4c5a9e390ca57b86b14e51e0c4dd5f7368dee4a20f210403c25b3c617ff0"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.338896 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dwhls" event={"ID":"69ac761e-0a13-405b-be5a-60ad535b77a7","Type":"ContainerStarted","Data":"7c73118fb8cfaf84a9942e5dcf8d4f5b367fe6928b839eb9675f6afa89835958"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.345190 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zjjnf" event={"ID":"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb","Type":"ContainerStarted","Data":"4f46be2ad28273415132ddc59b7d8ad159e695efb3ac4b6500af34424f6776b8"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.345327 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zjjnf" event={"ID":"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb","Type":"ContainerStarted","Data":"c133faedbb7deeb82849e6c0e23d11bf6c53a50af0010958394bdc2aa347cb24"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.346516 4936 generic.go:334] "Generic (PLEG): container finished" podID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerID="00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3" exitCode=0 Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.346667 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.346705 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerStarted","Data":"c96029774b5855d58e60bfc1f9bcfbf99acffaec31a6a5f3bfb970ad9a0520e4"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.347094 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-7vbr8" podStartSLOduration=38.347070976 podStartE2EDuration="38.347070976s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:29.344954568 +0000 UTC m=+100.291322383" watchObservedRunningTime="2026-03-20 16:02:29.347070976 +0000 UTC m=+100.293438791" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.362627 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"0b3a961bc4c3f4a208ccba10e68d2dd183fbdd6e7d0527f70542d266e61c0026"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.362694 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"548766dd512244399f51b55b943c4d13acad4bf13ce80ec19f3b99a712997270"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.362706 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"223e93494043318c27f673bf9a39b69cda1b1d1c5decca43129aa4d7841c5b8c"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.388490 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.388534 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.388562 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.388583 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.388625 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:29Z","lastTransitionTime":"2026-03-20T16:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.430203 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podStartSLOduration=38.43017618 podStartE2EDuration="38.43017618s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:29.429437829 +0000 UTC m=+100.375805644" watchObservedRunningTime="2026-03-20 16:02:29.43017618 +0000 UTC m=+100.376544005" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.451107 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-zjjnf" podStartSLOduration=38.451084522 podStartE2EDuration="38.451084522s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:29.450161246 +0000 UTC m=+100.396529061" watchObservedRunningTime="2026-03-20 16:02:29.451084522 +0000 UTC m=+100.397452337" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.491214 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.491329 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.491350 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.491377 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.491395 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:29Z","lastTransitionTime":"2026-03-20T16:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.595712 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.595764 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.595773 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.595791 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.595834 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:29Z","lastTransitionTime":"2026-03-20T16:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.601640 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.601721 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.601755 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.601780 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.601809 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.601915 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.601936 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.601946 4936 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.601987 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:33.60197359 +0000 UTC m=+104.548341405 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.602451 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.602481 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.602493 4936 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.602565 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:33.602527505 +0000 UTC m=+104.548895470 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.602652 4936 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.602741 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:33.602664848 +0000 UTC m=+104.549032663 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.602835 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:33.602796582 +0000 UTC m=+104.549164587 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.603036 4936 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.603116 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:33.60310307 +0000 UTC m=+104.549471045 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.698737 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.698797 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.698809 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.698828 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.698841 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:29Z","lastTransitionTime":"2026-03-20T16:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.703254 4936 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.703330 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.703437 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs podName:b65a2c6b-8b39-4742-bd96-6c4cbdc768ac nodeName:}" failed. No retries permitted until 2026-03-20 16:02:30.703398824 +0000 UTC m=+101.649766639 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs") pod "network-metrics-daemon-7jnpn" (UID: "b65a2c6b-8b39-4742-bd96-6c4cbdc768ac") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.801820 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.801876 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.801887 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.801908 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.801920 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:29Z","lastTransitionTime":"2026-03-20T16:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.853240 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.853284 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.854521 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 16:02:29 crc kubenswrapper[4936]: E0320 16:02:29.854701 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.905166 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.905203 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.905211 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.905229 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:29 crc kubenswrapper[4936]: I0320 16:02:29.905239 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:29Z","lastTransitionTime":"2026-03-20T16:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.008668 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.008719 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.008730 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.008750 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.008763 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.111837 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.112252 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.112266 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.112297 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.112326 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.216334 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.216391 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.216416 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.216448 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.216465 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.320733 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.320823 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.320844 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.320882 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.320909 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.373441 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerStarted","Data":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.373532 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerStarted","Data":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.373577 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerStarted","Data":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.373598 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerStarted","Data":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.376062 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerDied","Data":"49e88b71c957a8ebb056938c31f940d0e00c7629c646b8bfd2d4b1a0ce3b5232"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.376064 4936 generic.go:334] "Generic (PLEG): container finished" podID="e3d9dfab-9151-432d-ab1b-a78c76959b70" containerID="49e88b71c957a8ebb056938c31f940d0e00c7629c646b8bfd2d4b1a0ce3b5232" exitCode=0 Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.379119 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dwhls" event={"ID":"69ac761e-0a13-405b-be5a-60ad535b77a7","Type":"ContainerStarted","Data":"3e45bcc382e81ed4e0de76313ee9eadd8a73fbd7322daee473d0eb7ddbf91bb3"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.389715 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" event={"ID":"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c","Type":"ContainerStarted","Data":"2a26929988cfb0c181aa05875aff75f44cfc4c243e353de4993d351c187c41a3"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.390824 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" event={"ID":"6c4d53e5-257f-435a-88f1-9c2e5ad1f10c","Type":"ContainerStarted","Data":"f7cbc0dd0057e257bbeb1cd04bf4c3e1d9a0aa81fe35c49ee56b2967c5da8bb3"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.428377 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.428442 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.428461 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.428487 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.428524 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.430556 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dwhls" podStartSLOduration=39.430512075 podStartE2EDuration="39.430512075s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:30.429055585 +0000 UTC m=+101.375423410" watchObservedRunningTime="2026-03-20 16:02:30.430512075 +0000 UTC m=+101.376879890" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.459683 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpbcc" podStartSLOduration=39.459660763 podStartE2EDuration="39.459660763s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:30.458693366 +0000 UTC m=+101.405061181" watchObservedRunningTime="2026-03-20 16:02:30.459660763 +0000 UTC m=+101.406028578" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.532000 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.532460 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.532477 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.532502 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.532516 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.636778 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.636820 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.636831 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.636849 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.636863 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.713958 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:30 crc kubenswrapper[4936]: E0320 16:02:30.714231 4936 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:30 crc kubenswrapper[4936]: E0320 16:02:30.714358 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs podName:b65a2c6b-8b39-4742-bd96-6c4cbdc768ac nodeName:}" failed. No retries permitted until 2026-03-20 16:02:32.71432956 +0000 UTC m=+103.660697375 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs") pod "network-metrics-daemon-7jnpn" (UID: "b65a2c6b-8b39-4742-bd96-6c4cbdc768ac") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.739779 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.739823 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.739834 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.739852 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.739864 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.842667 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.842730 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.842743 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.842764 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.842778 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.853381 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.853471 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:30 crc kubenswrapper[4936]: E0320 16:02:30.853516 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 16:02:30 crc kubenswrapper[4936]: E0320 16:02:30.853693 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7jnpn" podUID="b65a2c6b-8b39-4742-bd96-6c4cbdc768ac" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.945442 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.945893 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.945908 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.945929 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:30 crc kubenswrapper[4936]: I0320 16:02:30.945946 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:30Z","lastTransitionTime":"2026-03-20T16:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.048529 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.048587 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.048597 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.048614 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.048627 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.151678 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.151724 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.151737 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.151756 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.151769 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.255614 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.255793 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.256010 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.256047 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.256073 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.359837 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.359895 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.359911 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.359935 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.359956 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.397894 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerStarted","Data":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.397974 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerStarted","Data":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.400679 4936 generic.go:334] "Generic (PLEG): container finished" podID="e3d9dfab-9151-432d-ab1b-a78c76959b70" containerID="727811464cb220ce9a2606809b8902b4c96ab1f648f62da4eecbd1ca2ca65c22" exitCode=0 Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.400745 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerDied","Data":"727811464cb220ce9a2606809b8902b4c96ab1f648f62da4eecbd1ca2ca65c22"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.463825 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.464055 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.464067 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.464088 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.464103 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.568026 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.568103 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.568125 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.568155 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.568176 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.670736 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.670822 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.670844 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.670874 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.670895 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.774139 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.774191 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.774207 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.774228 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.774246 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.853876 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.853935 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:31 crc kubenswrapper[4936]: E0320 16:02:31.854042 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 16:02:31 crc kubenswrapper[4936]: E0320 16:02:31.854211 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.877389 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.877443 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.877460 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.877487 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.877507 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.980771 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.980825 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.980838 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.980859 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:31 crc kubenswrapper[4936]: I0320 16:02:31.980873 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:31Z","lastTransitionTime":"2026-03-20T16:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.084981 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.085052 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.085071 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.085099 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.085120 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:32Z","lastTransitionTime":"2026-03-20T16:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.194872 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.194926 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.194946 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.194986 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.195005 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:32Z","lastTransitionTime":"2026-03-20T16:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.298065 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.298108 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.298118 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.298135 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.298149 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:32Z","lastTransitionTime":"2026-03-20T16:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.401045 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.401119 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.401142 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.401177 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.401201 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:32Z","lastTransitionTime":"2026-03-20T16:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.407819 4936 generic.go:334] "Generic (PLEG): container finished" podID="e3d9dfab-9151-432d-ab1b-a78c76959b70" containerID="cb2d95c65d04cfc120823bf52b8224ee9372928721feacf8ce9f833e8921596b" exitCode=0 Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.408361 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerDied","Data":"cb2d95c65d04cfc120823bf52b8224ee9372928721feacf8ce9f833e8921596b"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.511652 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.512008 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.512059 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.512112 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.512136 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:32Z","lastTransitionTime":"2026-03-20T16:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.616584 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.616668 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.616688 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.616718 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.616740 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:32Z","lastTransitionTime":"2026-03-20T16:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.720910 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.720963 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.720978 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.720999 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.721011 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:32Z","lastTransitionTime":"2026-03-20T16:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.738503 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:32 crc kubenswrapper[4936]: E0320 16:02:32.738769 4936 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:32 crc kubenswrapper[4936]: E0320 16:02:32.738886 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs podName:b65a2c6b-8b39-4742-bd96-6c4cbdc768ac nodeName:}" failed. No retries permitted until 2026-03-20 16:02:36.738850645 +0000 UTC m=+107.685218480 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs") pod "network-metrics-daemon-7jnpn" (UID: "b65a2c6b-8b39-4742-bd96-6c4cbdc768ac") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.823795 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.823832 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.823841 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.823858 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.823869 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:32Z","lastTransitionTime":"2026-03-20T16:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.853872 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:32 crc kubenswrapper[4936]: E0320 16:02:32.854063 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7jnpn" podUID="b65a2c6b-8b39-4742-bd96-6c4cbdc768ac" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.854598 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:32 crc kubenswrapper[4936]: E0320 16:02:32.854676 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.875297 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.927051 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.927127 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.927145 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.927179 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:32 crc kubenswrapper[4936]: I0320 16:02:32.927200 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:32Z","lastTransitionTime":"2026-03-20T16:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.030655 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.030703 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.030714 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.030735 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.030747 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.134090 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.134148 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.134158 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.134175 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.134186 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.237030 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.237077 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.237089 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.237113 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.237127 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.340557 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.340619 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.340631 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.340652 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.340665 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.418277 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerStarted","Data":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.422263 4936 generic.go:334] "Generic (PLEG): container finished" podID="e3d9dfab-9151-432d-ab1b-a78c76959b70" containerID="3cf22011e605b8f13a9aeea40772f52217f53919606b15f1cad60e7fcb694107" exitCode=0 Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.422353 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerDied","Data":"3cf22011e605b8f13a9aeea40772f52217f53919606b15f1cad60e7fcb694107"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.443749 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.443811 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.443832 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.443861 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.443884 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.471033 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.471002565 podStartE2EDuration="1.471002565s" podCreationTimestamp="2026-03-20 16:02:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:33.470044238 +0000 UTC m=+104.416412093" watchObservedRunningTime="2026-03-20 16:02:33.471002565 +0000 UTC m=+104.417370390" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.547137 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.547176 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.547187 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.547206 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.547218 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.649379 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.649466 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.649509 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.649559 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:41.649507178 +0000 UTC m=+112.595875163 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.649620 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.649690 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.649707 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.649716 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.650099 4936 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.650254 4936 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.650356 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.650386 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:41.650337841 +0000 UTC m=+112.596705676 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.650394 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.650401 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.650423 4936 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.650431 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:41.650410223 +0000 UTC m=+112.596778058 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.650453 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.650475 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.650501 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:41.650484585 +0000 UTC m=+112.596852400 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.650507 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.650533 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.654818 4936 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.654922 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:41.654896785 +0000 UTC m=+112.601264600 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.753849 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.753928 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.753956 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.753985 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.754004 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.853790 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.853956 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.854022 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:33 crc kubenswrapper[4936]: E0320 16:02:33.854736 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.857077 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.857140 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.857154 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.857175 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.857188 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.961393 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.961459 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.961478 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.961496 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:33 crc kubenswrapper[4936]: I0320 16:02:33.961516 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:33Z","lastTransitionTime":"2026-03-20T16:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.063990 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.064046 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.064057 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.064083 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.064107 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.166351 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.166408 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.166421 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.166438 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.166448 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.270046 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.270104 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.270116 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.270140 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.270158 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.373378 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.373438 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.373453 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.373479 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.373495 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.430725 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerStarted","Data":"af22dcd4f8a889ec25d4ae33e111edd8f79c991e9f11199d3e8cf57d53fb0024"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.476938 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.476993 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.477014 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.477041 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.477061 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.579949 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.580020 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.580044 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.580079 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.580104 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.683677 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.683739 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.683753 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.683779 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.683793 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.788259 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.788347 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.788367 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.788393 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.788413 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.853910 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:34 crc kubenswrapper[4936]: E0320 16:02:34.854134 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7jnpn" podUID="b65a2c6b-8b39-4742-bd96-6c4cbdc768ac" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.854760 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:34 crc kubenswrapper[4936]: E0320 16:02:34.854867 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.891022 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.891348 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.891358 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.891376 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.891425 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.994236 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.994294 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.994307 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.994327 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:34 crc kubenswrapper[4936]: I0320 16:02:34.994339 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:34Z","lastTransitionTime":"2026-03-20T16:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.097179 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.097249 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.097271 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.097304 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.097327 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:35Z","lastTransitionTime":"2026-03-20T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.200726 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.200789 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.200804 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.200830 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.200845 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:35Z","lastTransitionTime":"2026-03-20T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.304071 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.304147 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.304167 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.304195 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.304220 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:35Z","lastTransitionTime":"2026-03-20T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.408008 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.408054 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.408067 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.408089 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.408100 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:35Z","lastTransitionTime":"2026-03-20T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.441415 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerStarted","Data":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.441977 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.442005 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.446464 4936 generic.go:334] "Generic (PLEG): container finished" podID="e3d9dfab-9151-432d-ab1b-a78c76959b70" containerID="af22dcd4f8a889ec25d4ae33e111edd8f79c991e9f11199d3e8cf57d53fb0024" exitCode=0 Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.446539 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerDied","Data":"af22dcd4f8a889ec25d4ae33e111edd8f79c991e9f11199d3e8cf57d53fb0024"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.533095 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" podStartSLOduration=44.533076706 podStartE2EDuration="44.533076706s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:35.53286873 +0000 UTC m=+106.479236575" watchObservedRunningTime="2026-03-20 16:02:35.533076706 +0000 UTC m=+106.479444521" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.540708 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.540743 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.540757 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.540775 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.540790 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:35Z","lastTransitionTime":"2026-03-20T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.541920 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.643460 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.643504 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.643516 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.643533 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.643557 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:35Z","lastTransitionTime":"2026-03-20T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.746300 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.746346 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.746359 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.746380 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.746396 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:35Z","lastTransitionTime":"2026-03-20T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.849406 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.849459 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.849470 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.849492 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.849502 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:35Z","lastTransitionTime":"2026-03-20T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.853710 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.853750 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:35 crc kubenswrapper[4936]: E0320 16:02:35.853846 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 16:02:35 crc kubenswrapper[4936]: E0320 16:02:35.853984 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.954259 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.954313 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.954328 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.954351 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:35 crc kubenswrapper[4936]: I0320 16:02:35.954365 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:35Z","lastTransitionTime":"2026-03-20T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.057694 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.057738 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.057747 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.057769 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.057778 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.161378 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.161438 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.161452 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.161476 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.161490 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.264412 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.264467 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.264479 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.264499 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.264515 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.368845 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.368953 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.368975 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.369008 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.369031 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.457249 4936 generic.go:334] "Generic (PLEG): container finished" podID="e3d9dfab-9151-432d-ab1b-a78c76959b70" containerID="392dee367aaf6da0f74e98536cdcf8a712aa7357dbf21f8bc8a830225d7b6026" exitCode=0 Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.458072 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerDied","Data":"392dee367aaf6da0f74e98536cdcf8a712aa7357dbf21f8bc8a830225d7b6026"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.458948 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.472097 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.472149 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.472162 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.472182 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.472197 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.500804 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.575927 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.575986 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.576003 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.576028 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.576047 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.679024 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.679070 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.679082 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.679101 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.679116 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.748659 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:36 crc kubenswrapper[4936]: E0320 16:02:36.749017 4936 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:36 crc kubenswrapper[4936]: E0320 16:02:36.749156 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs podName:b65a2c6b-8b39-4742-bd96-6c4cbdc768ac nodeName:}" failed. No retries permitted until 2026-03-20 16:02:44.749118473 +0000 UTC m=+115.695486478 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs") pod "network-metrics-daemon-7jnpn" (UID: "b65a2c6b-8b39-4742-bd96-6c4cbdc768ac") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.781935 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.781990 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.782010 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.782035 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.782049 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.853667 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.853805 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:36 crc kubenswrapper[4936]: E0320 16:02:36.853817 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7jnpn" podUID="b65a2c6b-8b39-4742-bd96-6c4cbdc768ac" Mar 20 16:02:36 crc kubenswrapper[4936]: E0320 16:02:36.854068 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.884854 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.884913 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.884930 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.884953 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.884967 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.987275 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.987324 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.987336 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.987354 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:36 crc kubenswrapper[4936]: I0320 16:02:36.987369 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:36Z","lastTransitionTime":"2026-03-20T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.090024 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.090076 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.090086 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.090104 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.090115 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:37Z","lastTransitionTime":"2026-03-20T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.193684 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.193737 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.193755 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.193778 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.193795 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:37Z","lastTransitionTime":"2026-03-20T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.312088 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7jnpn"] Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.312648 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.312688 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.312704 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.312731 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.312747 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:37Z","lastTransitionTime":"2026-03-20T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.416557 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.416603 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.416620 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.416644 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.416659 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:37Z","lastTransitionTime":"2026-03-20T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.465140 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:37 crc kubenswrapper[4936]: E0320 16:02:37.465302 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7jnpn" podUID="b65a2c6b-8b39-4742-bd96-6c4cbdc768ac" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.465874 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sp67w" event={"ID":"e3d9dfab-9151-432d-ab1b-a78c76959b70","Type":"ContainerStarted","Data":"cf75e7b21fd9685c48061367c2e2e6fe49fc3d1560558755e558673061a68641"} Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.496980 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-sp67w" podStartSLOduration=46.496776807 podStartE2EDuration="46.496776807s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:37.494372651 +0000 UTC m=+108.440740486" watchObservedRunningTime="2026-03-20 16:02:37.496776807 +0000 UTC m=+108.443144642" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.519352 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.519402 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.519413 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.519427 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.519438 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:37Z","lastTransitionTime":"2026-03-20T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.558886 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.558956 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.558972 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.558989 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.559002 4936 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T16:02:37Z","lastTransitionTime":"2026-03-20T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.604777 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc"] Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.605356 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.609037 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.609434 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.609458 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.609462 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.759537 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/981fbe70-eddb-4cec-8272-e928d325db1e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.759609 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/981fbe70-eddb-4cec-8272-e928d325db1e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.759841 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/981fbe70-eddb-4cec-8272-e928d325db1e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.759951 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/981fbe70-eddb-4cec-8272-e928d325db1e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.760031 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/981fbe70-eddb-4cec-8272-e928d325db1e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.836555 4936 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.845936 4936 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.854031 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:37 crc kubenswrapper[4936]: E0320 16:02:37.854224 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.854045 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:37 crc kubenswrapper[4936]: E0320 16:02:37.854447 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.861086 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/981fbe70-eddb-4cec-8272-e928d325db1e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.861155 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/981fbe70-eddb-4cec-8272-e928d325db1e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.861189 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/981fbe70-eddb-4cec-8272-e928d325db1e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.861232 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/981fbe70-eddb-4cec-8272-e928d325db1e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.861186 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/981fbe70-eddb-4cec-8272-e928d325db1e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.861265 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/981fbe70-eddb-4cec-8272-e928d325db1e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.861338 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/981fbe70-eddb-4cec-8272-e928d325db1e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.862944 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/981fbe70-eddb-4cec-8272-e928d325db1e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.871270 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/981fbe70-eddb-4cec-8272-e928d325db1e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.878238 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/981fbe70-eddb-4cec-8272-e928d325db1e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dlhhc\" (UID: \"981fbe70-eddb-4cec-8272-e928d325db1e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: I0320 16:02:37.926841 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" Mar 20 16:02:37 crc kubenswrapper[4936]: W0320 16:02:37.946784 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod981fbe70_eddb_4cec_8272_e928d325db1e.slice/crio-a3b019ae1c7712ce8224caefdc000c389e88b43dc3f1b6712e40a8728a1d5275 WatchSource:0}: Error finding container a3b019ae1c7712ce8224caefdc000c389e88b43dc3f1b6712e40a8728a1d5275: Status 404 returned error can't find the container with id a3b019ae1c7712ce8224caefdc000c389e88b43dc3f1b6712e40a8728a1d5275 Mar 20 16:02:38 crc kubenswrapper[4936]: I0320 16:02:38.469860 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" event={"ID":"981fbe70-eddb-4cec-8272-e928d325db1e","Type":"ContainerStarted","Data":"cc7d9f8b013091439a537494c392f6fb8b43d4f3b8f4bbad3310e9cf2b213635"} Mar 20 16:02:38 crc kubenswrapper[4936]: I0320 16:02:38.470289 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" event={"ID":"981fbe70-eddb-4cec-8272-e928d325db1e","Type":"ContainerStarted","Data":"a3b019ae1c7712ce8224caefdc000c389e88b43dc3f1b6712e40a8728a1d5275"} Mar 20 16:02:38 crc kubenswrapper[4936]: I0320 16:02:38.491847 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dlhhc" podStartSLOduration=47.491826028 podStartE2EDuration="47.491826028s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:38.490845281 +0000 UTC m=+109.437213096" watchObservedRunningTime="2026-03-20 16:02:38.491826028 +0000 UTC m=+109.438193843" Mar 20 16:02:38 crc kubenswrapper[4936]: I0320 16:02:38.854198 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:38 crc kubenswrapper[4936]: E0320 16:02:38.854369 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 16:02:38 crc kubenswrapper[4936]: I0320 16:02:38.972794 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 16:02:38 crc kubenswrapper[4936]: I0320 16:02:38.972829 4936 scope.go:117] "RemoveContainer" containerID="034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594" Mar 20 16:02:38 crc kubenswrapper[4936]: E0320 16:02:38.973040 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:02:39 crc kubenswrapper[4936]: I0320 16:02:39.474832 4936 scope.go:117] "RemoveContainer" containerID="034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594" Mar 20 16:02:39 crc kubenswrapper[4936]: E0320 16:02:39.475037 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 16:02:39 crc kubenswrapper[4936]: I0320 16:02:39.853687 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:39 crc kubenswrapper[4936]: I0320 16:02:39.853706 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:39 crc kubenswrapper[4936]: I0320 16:02:39.853733 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:39 crc kubenswrapper[4936]: E0320 16:02:39.854947 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 16:02:39 crc kubenswrapper[4936]: E0320 16:02:39.855023 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 16:02:39 crc kubenswrapper[4936]: E0320 16:02:39.855084 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7jnpn" podUID="b65a2c6b-8b39-4742-bd96-6c4cbdc768ac" Mar 20 16:02:40 crc kubenswrapper[4936]: I0320 16:02:40.853188 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:40 crc kubenswrapper[4936]: E0320 16:02:40.853394 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.255689 4936 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.255967 4936 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.322715 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4p8zd"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.323715 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.324437 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ssr5n"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.325505 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.327910 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.328297 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.329817 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dkllp"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.330202 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.331540 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jhqvm"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.332245 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.335625 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jgfjb"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.336233 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.354308 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.354327 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.354737 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.354940 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.355600 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.355954 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.356024 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.356330 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.359866 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.360230 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.360917 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.361645 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.362059 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.362069 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.362202 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.362266 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.362383 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.362511 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.362612 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.362617 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.362745 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.363125 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.366024 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.366123 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.366162 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.367039 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.367276 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.367420 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.368537 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.372457 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.372674 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.372762 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.372805 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.373063 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.373219 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.373363 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.373522 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.373775 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.384785 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.385132 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.385317 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.387461 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.387882 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.387981 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.388817 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.388873 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.389048 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.389153 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.389213 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.389328 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.389605 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.389752 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.389872 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.389999 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.390015 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.390603 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.410736 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411021 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wstk\" (UniqueName: \"kubernetes.io/projected/d815687f-6e2a-44c8-bf71-c9718db7ba67-kube-api-access-6wstk\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411082 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtxgl\" (UniqueName: \"kubernetes.io/projected/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-kube-api-access-qtxgl\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411113 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d815687f-6e2a-44c8-bf71-c9718db7ba67-images\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411132 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d815687f-6e2a-44c8-bf71-c9718db7ba67-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411158 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411177 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d815687f-6e2a-44c8-bf71-c9718db7ba67-config\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411237 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-config\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411256 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-client-ca\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411274 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-serving-cert\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411396 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-dzgl2"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.411673 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.412162 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.412224 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.412977 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dzgl2" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.416518 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.417329 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.418799 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.418988 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.419706 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.419791 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.419884 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.420019 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.420255 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.420348 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.420481 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.422227 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.422562 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.422763 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.425489 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.428147 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-6q5fl"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.428876 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q6cx6"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.429219 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-m97k8"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.429527 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.429732 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.430097 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.430889 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.433769 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.435979 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.439020 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xtqwz"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.450609 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.451304 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.451924 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.452362 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.455565 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.456171 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.459289 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.459463 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.459764 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.459951 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.460101 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.460868 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.461108 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.461314 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.462927 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.463861 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.464246 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.464499 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.465930 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.466655 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.467792 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.467826 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.468143 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.470613 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.473051 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8tlt4"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.473253 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.473650 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.473880 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.474118 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.474129 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.474206 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.474425 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.475018 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.475090 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.474582 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.474675 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.475669 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.476290 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.505518 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j54jw"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.506105 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.506712 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.506863 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.507367 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.507600 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.510103 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.510819 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.511685 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.511911 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d71d251-484c-4198-b697-298f2d92031f-config\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.511945 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-image-import-ca\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.511965 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.511988 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtxgl\" (UniqueName: \"kubernetes.io/projected/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-kube-api-access-qtxgl\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512009 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d815687f-6e2a-44c8-bf71-c9718db7ba67-images\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512029 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d815687f-6e2a-44c8-bf71-c9718db7ba67-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512048 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8498b530-52d4-4ca5-88ca-fda75dfbb18a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512068 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7175b7-4b92-49c8-98d1-b2366c099ced-serving-cert\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512088 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7e7175b7-4b92-49c8-98d1-b2366c099ced-audit-policies\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512109 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8498b530-52d4-4ca5-88ca-fda75dfbb18a-serving-cert\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512146 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512170 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512201 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512222 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512239 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512258 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-trusted-ca-bundle\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512273 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-oauth-serving-cert\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512292 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frvx8\" (UniqueName: \"kubernetes.io/projected/bf0f9e97-bddb-4054-99dd-d10428233e9f-kube-api-access-frvx8\") pod \"downloads-7954f5f757-dzgl2\" (UID: \"bf0f9e97-bddb-4054-99dd-d10428233e9f\") " pod="openshift-console/downloads-7954f5f757-dzgl2" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512309 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg5hx\" (UniqueName: \"kubernetes.io/projected/9333cf9d-b80d-4242-a940-c92ed3b593d1-kube-api-access-rg5hx\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512325 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-oauth-config\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512342 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5ghv\" (UniqueName: \"kubernetes.io/projected/0f7a4f62-f974-487f-90d0-f45d16638c4c-kube-api-access-f5ghv\") pod \"cluster-samples-operator-665b6dd947-p8jqw\" (UID: \"0f7a4f62-f974-487f-90d0-f45d16638c4c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512359 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7e7175b7-4b92-49c8-98d1-b2366c099ced-encryption-config\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512377 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5llft\" (UniqueName: \"kubernetes.io/projected/7d71d251-484c-4198-b697-298f2d92031f-kube-api-access-5llft\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512393 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-etcd-serving-ca\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512411 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-config\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512432 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-client-ca\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512450 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpxkx\" (UniqueName: \"kubernetes.io/projected/c52c0329-ca68-4d95-8341-ab8191c3a186-kube-api-access-wpxkx\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512468 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf423681-0ee6-4d44-bbed-e1447eaa7edd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-k98jk\" (UID: \"cf423681-0ee6-4d44-bbed-e1447eaa7edd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512487 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wstk\" (UniqueName: \"kubernetes.io/projected/d815687f-6e2a-44c8-bf71-c9718db7ba67-kube-api-access-6wstk\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512507 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-config\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512524 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7e7175b7-4b92-49c8-98d1-b2366c099ced-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512557 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48jmz\" (UniqueName: \"kubernetes.io/projected/8498b530-52d4-4ca5-88ca-fda75dfbb18a-kube-api-access-48jmz\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512577 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-serving-cert\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512600 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj2cd\" (UniqueName: \"kubernetes.io/projected/7e7175b7-4b92-49c8-98d1-b2366c099ced-kube-api-access-nj2cd\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512618 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7d71d251-484c-4198-b697-298f2d92031f-auth-proxy-config\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512638 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512658 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512675 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512693 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7e7175b7-4b92-49c8-98d1-b2366c099ced-audit-dir\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512710 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdcq2\" (UniqueName: \"kubernetes.io/projected/cf423681-0ee6-4d44-bbed-e1447eaa7edd-kube-api-access-zdcq2\") pod \"openshift-apiserver-operator-796bbdcf4f-k98jk\" (UID: \"cf423681-0ee6-4d44-bbed-e1447eaa7edd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512731 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d815687f-6e2a-44c8-bf71-c9718db7ba67-config\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512754 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8498b530-52d4-4ca5-88ca-fda75dfbb18a-service-ca-bundle\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512776 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512796 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4gzw\" (UniqueName: \"kubernetes.io/projected/6376c9a4-9209-4f0d-835d-7db09de15649-kube-api-access-j4gzw\") pod \"openshift-config-operator-7777fb866f-m97k8\" (UID: \"6376c9a4-9209-4f0d-835d-7db09de15649\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512824 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/151dc193-527a-4506-bcb9-170641b9ea3f-audit-dir\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512846 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512870 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-console-config\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512890 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-audit\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512907 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-policies\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512939 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512955 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7d71d251-484c-4198-b697-298f2d92031f-machine-approver-tls\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.512974 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf423681-0ee6-4d44-bbed-e1447eaa7edd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-k98jk\" (UID: \"cf423681-0ee6-4d44-bbed-e1447eaa7edd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513006 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6376c9a4-9209-4f0d-835d-7db09de15649-available-featuregates\") pod \"openshift-config-operator-7777fb866f-m97k8\" (UID: \"6376c9a4-9209-4f0d-835d-7db09de15649\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513026 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/151dc193-527a-4506-bcb9-170641b9ea3f-node-pullsecrets\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513045 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc7q8\" (UniqueName: \"kubernetes.io/projected/151dc193-527a-4506-bcb9-170641b9ea3f-kube-api-access-zc7q8\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513061 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-service-ca\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513076 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7e7175b7-4b92-49c8-98d1-b2366c099ced-etcd-client\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513092 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/151dc193-527a-4506-bcb9-170641b9ea3f-encryption-config\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513122 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0f7a4f62-f974-487f-90d0-f45d16638c4c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p8jqw\" (UID: \"0f7a4f62-f974-487f-90d0-f45d16638c4c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513142 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/151dc193-527a-4506-bcb9-170641b9ea3f-serving-cert\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513163 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1f1994bb-a56f-4e25-b6d8-7bcc2113de10-metrics-tls\") pod \"dns-operator-744455d44c-q6cx6\" (UID: \"1f1994bb-a56f-4e25-b6d8-7bcc2113de10\") " pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513179 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/151dc193-527a-4506-bcb9-170641b9ea3f-etcd-client\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513200 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-serving-cert\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513222 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8498b530-52d4-4ca5-88ca-fda75dfbb18a-config\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513247 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-dir\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513282 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e7175b7-4b92-49c8-98d1-b2366c099ced-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513325 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6376c9a4-9209-4f0d-835d-7db09de15649-serving-cert\") pod \"openshift-config-operator-7777fb866f-m97k8\" (UID: \"6376c9a4-9209-4f0d-835d-7db09de15649\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.513087 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d815687f-6e2a-44c8-bf71-c9718db7ba67-images\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.514403 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.514451 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwssr\" (UniqueName: \"kubernetes.io/projected/1f1994bb-a56f-4e25-b6d8-7bcc2113de10-kube-api-access-bwssr\") pod \"dns-operator-744455d44c-q6cx6\" (UID: \"1f1994bb-a56f-4e25-b6d8-7bcc2113de10\") " pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.514479 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.514693 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d815687f-6e2a-44c8-bf71-c9718db7ba67-config\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.514993 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.515248 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-client-ca\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.515575 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4p8zd"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.516228 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-config\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.517703 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d815687f-6e2a-44c8-bf71-c9718db7ba67-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.517871 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.517927 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.518565 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.518641 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.518637 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.518670 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.518716 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.520279 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-serving-cert\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.520338 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.520952 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.523471 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ssr5n"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.524576 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.525416 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.525662 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xlnmq"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.526216 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.526308 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.527252 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-m5rqk"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.527578 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.528184 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.528682 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r4w72"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.529346 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.530736 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.531117 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.534638 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.535041 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.538600 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.539070 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.541587 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.542334 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.546491 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.550946 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.551732 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.551945 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.553631 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.554624 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.557480 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.560224 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.562980 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8v7wv"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.565284 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-9ltqh"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.565708 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.566708 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.568256 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.571932 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dkllp"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.571986 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jhqvm"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.572157 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.572329 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.583258 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jgfjb"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.585067 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8tlt4"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.586250 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.586422 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r4w72"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.588622 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.588646 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dzgl2"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.589231 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.590343 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q6cx6"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.591441 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.593026 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-m97k8"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.593795 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.595306 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.595974 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xtqwz"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.598132 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.599070 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-5nbl4"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.600556 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6q5fl"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.600778 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.602203 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.603130 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qbm7q"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.603841 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.603970 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qbm7q" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.606049 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.606186 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.606220 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.607223 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.608488 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.609465 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.610481 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.611585 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j54jw"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.612697 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.613818 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.614599 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xlnmq"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.615306 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgp2g\" (UniqueName: \"kubernetes.io/projected/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-kube-api-access-zgp2g\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.615343 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1f1994bb-a56f-4e25-b6d8-7bcc2113de10-metrics-tls\") pod \"dns-operator-744455d44c-q6cx6\" (UID: \"1f1994bb-a56f-4e25-b6d8-7bcc2113de10\") " pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.615368 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c6hh\" (UniqueName: \"kubernetes.io/projected/ab39813c-c2d7-432c-aa05-7b33c1e86fac-kube-api-access-5c6hh\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.615389 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/242b082c-0e13-4b74-830e-0bb1008b5c0a-profile-collector-cert\") pod \"catalog-operator-68c6474976-ccjqw\" (UID: \"242b082c-0e13-4b74-830e-0bb1008b5c0a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.615429 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e7175b7-4b92-49c8-98d1-b2366c099ced-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.615601 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8v7wv"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.615650 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6376c9a4-9209-4f0d-835d-7db09de15649-serving-cert\") pod \"openshift-config-operator-7777fb866f-m97k8\" (UID: \"6376c9a4-9209-4f0d-835d-7db09de15649\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616005 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616067 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e7175b7-4b92-49c8-98d1-b2366c099ced-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616145 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwssr\" (UniqueName: \"kubernetes.io/projected/1f1994bb-a56f-4e25-b6d8-7bcc2113de10-kube-api-access-bwssr\") pod \"dns-operator-744455d44c-q6cx6\" (UID: \"1f1994bb-a56f-4e25-b6d8-7bcc2113de10\") " pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616225 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f8268de-4dd1-4521-9e5a-b4679996b511-serving-cert\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616300 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-image-import-ca\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616447 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616532 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616532 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d71d251-484c-4198-b697-298f2d92031f-config\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616647 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd7j2\" (UniqueName: \"kubernetes.io/projected/32e4b213-a66b-4e97-83b1-cac96a741b9a-kube-api-access-pd7j2\") pod \"openshift-controller-manager-operator-756b6f6bc6-64r5t\" (UID: \"32e4b213-a66b-4e97-83b1-cac96a741b9a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616711 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/967b1700-c689-4195-8dcd-4177b702d066-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r4w72\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616756 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8498b530-52d4-4ca5-88ca-fda75dfbb18a-serving-cert\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616794 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616795 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616857 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r5jp\" (UniqueName: \"kubernetes.io/projected/fb1e412b-f423-4178-a306-b9201f089099-kube-api-access-5r5jp\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616895 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mgvd\" (UniqueName: \"kubernetes.io/projected/cd07f1f3-ea79-4612-af59-91151c9bb06f-kube-api-access-6mgvd\") pod \"control-plane-machine-set-operator-78cbb6b69f-vm2rn\" (UID: \"cd07f1f3-ea79-4612-af59-91151c9bb06f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616922 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/967b1700-c689-4195-8dcd-4177b702d066-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r4w72\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616952 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.616977 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg5hx\" (UniqueName: \"kubernetes.io/projected/9333cf9d-b80d-4242-a940-c92ed3b593d1-kube-api-access-rg5hx\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617004 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-oauth-config\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617031 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-trusted-ca-bundle\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617057 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frvx8\" (UniqueName: \"kubernetes.io/projected/bf0f9e97-bddb-4054-99dd-d10428233e9f-kube-api-access-frvx8\") pod \"downloads-7954f5f757-dzgl2\" (UID: \"bf0f9e97-bddb-4054-99dd-d10428233e9f\") " pod="openshift-console/downloads-7954f5f757-dzgl2" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617088 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5ghv\" (UniqueName: \"kubernetes.io/projected/0f7a4f62-f974-487f-90d0-f45d16638c4c-kube-api-access-f5ghv\") pod \"cluster-samples-operator-665b6dd947-p8jqw\" (UID: \"0f7a4f62-f974-487f-90d0-f45d16638c4c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617121 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7e7175b7-4b92-49c8-98d1-b2366c099ced-encryption-config\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617179 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9027bafb-1a39-4a53-9bcf-778b8859c192-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-58dfj\" (UID: \"9027bafb-1a39-4a53-9bcf-778b8859c192\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617207 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpxkx\" (UniqueName: \"kubernetes.io/projected/c52c0329-ca68-4d95-8341-ab8191c3a186-kube-api-access-wpxkx\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617247 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab39813c-c2d7-432c-aa05-7b33c1e86fac-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617274 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b6ffb849-1700-439d-b7d3-1c1540a48fc1-webhook-cert\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617304 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48jmz\" (UniqueName: \"kubernetes.io/projected/8498b530-52d4-4ca5-88ca-fda75dfbb18a-kube-api-access-48jmz\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617330 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9027bafb-1a39-4a53-9bcf-778b8859c192-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-58dfj\" (UID: \"9027bafb-1a39-4a53-9bcf-778b8859c192\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617357 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj2cd\" (UniqueName: \"kubernetes.io/projected/7e7175b7-4b92-49c8-98d1-b2366c099ced-kube-api-access-nj2cd\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617386 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab39813c-c2d7-432c-aa05-7b33c1e86fac-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617414 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-serving-cert\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617426 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qbm7q"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617444 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617561 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4gzw\" (UniqueName: \"kubernetes.io/projected/6376c9a4-9209-4f0d-835d-7db09de15649-kube-api-access-j4gzw\") pod \"openshift-config-operator-7777fb866f-m97k8\" (UID: \"6376c9a4-9209-4f0d-835d-7db09de15649\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617587 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ab39813c-c2d7-432c-aa05-7b33c1e86fac-ready\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617626 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8498b530-52d4-4ca5-88ca-fda75dfbb18a-service-ca-bundle\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617652 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/151dc193-527a-4506-bcb9-170641b9ea3f-audit-dir\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617671 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b6ffb849-1700-439d-b7d3-1c1540a48fc1-apiservice-cert\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617691 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-audit\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617716 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-policies\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617743 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6376c9a4-9209-4f0d-835d-7db09de15649-available-featuregates\") pod \"openshift-config-operator-7777fb866f-m97k8\" (UID: \"6376c9a4-9209-4f0d-835d-7db09de15649\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617760 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc7q8\" (UniqueName: \"kubernetes.io/projected/151dc193-527a-4506-bcb9-170641b9ea3f-kube-api-access-zc7q8\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617777 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx2rm\" (UniqueName: \"kubernetes.io/projected/4f8268de-4dd1-4521-9e5a-b4679996b511-kube-api-access-jx2rm\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617796 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-service-ca\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617799 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-image-import-ca\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617816 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf423681-0ee6-4d44-bbed-e1447eaa7edd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-k98jk\" (UID: \"cf423681-0ee6-4d44-bbed-e1447eaa7edd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617840 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb1e412b-f423-4178-a306-b9201f089099-images\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617860 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90dd52f0-03cd-46fd-90cd-1a480bae1280-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xxvbh\" (UID: \"90dd52f0-03cd-46fd-90cd-1a480bae1280\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617867 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617878 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0f7a4f62-f974-487f-90d0-f45d16638c4c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p8jqw\" (UID: \"0f7a4f62-f974-487f-90d0-f45d16638c4c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.617988 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/676cb1dc-3786-44f8-bf27-963da433b3c1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xlnmq\" (UID: \"676cb1dc-3786-44f8-bf27-963da433b3c1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618026 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fb1e412b-f423-4178-a306-b9201f089099-proxy-tls\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618051 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b6ffb849-1700-439d-b7d3-1c1540a48fc1-tmpfs\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618086 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/151dc193-527a-4506-bcb9-170641b9ea3f-serving-cert\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618109 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/151dc193-527a-4506-bcb9-170641b9ea3f-etcd-client\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618142 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8498b530-52d4-4ca5-88ca-fda75dfbb18a-config\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618168 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-dir\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618197 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjjfj\" (UniqueName: \"kubernetes.io/projected/967b1700-c689-4195-8dcd-4177b702d066-kube-api-access-mjjfj\") pod \"marketplace-operator-79b997595-r4w72\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618250 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfx84\" (UniqueName: \"kubernetes.io/projected/3a005177-344e-461b-b8ac-989f18ade02d-kube-api-access-bfx84\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618278 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fb1e412b-f423-4178-a306-b9201f089099-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618302 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtrk8\" (UniqueName: \"kubernetes.io/projected/242b082c-0e13-4b74-830e-0bb1008b5c0a-kube-api-access-gtrk8\") pod \"catalog-operator-68c6474976-ccjqw\" (UID: \"242b082c-0e13-4b74-830e-0bb1008b5c0a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.618327 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4f8268de-4dd1-4521-9e5a-b4679996b511-etcd-ca\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.619169 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d71d251-484c-4198-b697-298f2d92031f-config\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.619375 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-dir\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.619616 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1f1994bb-a56f-4e25-b6d8-7bcc2113de10-metrics-tls\") pod \"dns-operator-744455d44c-q6cx6\" (UID: \"1f1994bb-a56f-4e25-b6d8-7bcc2113de10\") " pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.619891 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5nbl4"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.620148 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621506 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621592 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f8268de-4dd1-4521-9e5a-b4679996b511-etcd-service-ca\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621624 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-policies\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621628 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8498b530-52d4-4ca5-88ca-fda75dfbb18a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621687 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621693 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7175b7-4b92-49c8-98d1-b2366c099ced-serving-cert\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621742 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7e7175b7-4b92-49c8-98d1-b2366c099ced-audit-policies\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621767 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm2jn\" (UniqueName: \"kubernetes.io/projected/b6ffb849-1700-439d-b7d3-1c1540a48fc1-kube-api-access-fm2jn\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621790 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/50acf6a8-8d00-4bd2-9a88-f85cd8b4b504-proxy-tls\") pod \"machine-config-controller-84d6567774-nnnkc\" (UID: \"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621845 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a005177-344e-461b-b8ac-989f18ade02d-serving-cert\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621872 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f8268de-4dd1-4521-9e5a-b4679996b511-config\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621902 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-oauth-serving-cert\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621924 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w6w4\" (UniqueName: \"kubernetes.io/projected/50acf6a8-8d00-4bd2-9a88-f85cd8b4b504-kube-api-access-4w6w4\") pod \"machine-config-controller-84d6567774-nnnkc\" (UID: \"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621947 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90dd52f0-03cd-46fd-90cd-1a480bae1280-config\") pod \"kube-controller-manager-operator-78b949d7b-xxvbh\" (UID: \"90dd52f0-03cd-46fd-90cd-1a480bae1280\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621973 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7bnp\" (UniqueName: \"kubernetes.io/projected/676cb1dc-3786-44f8-bf27-963da433b3c1-kube-api-access-p7bnp\") pod \"multus-admission-controller-857f4d67dd-xlnmq\" (UID: \"676cb1dc-3786-44f8-bf27-963da433b3c1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621996 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622017 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622066 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622087 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/242b082c-0e13-4b74-830e-0bb1008b5c0a-srv-cert\") pod \"catalog-operator-68c6474976-ccjqw\" (UID: \"242b082c-0e13-4b74-830e-0bb1008b5c0a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622110 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5llft\" (UniqueName: \"kubernetes.io/projected/7d71d251-484c-4198-b697-298f2d92031f-kube-api-access-5llft\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622131 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-etcd-serving-ca\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622140 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622152 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-787z7\" (UniqueName: \"kubernetes.io/projected/1cd73a43-1689-479b-bae3-6519bdd522b5-kube-api-access-787z7\") pod \"migrator-59844c95c7-42wg6\" (UID: \"1cd73a43-1689-479b-bae3-6519bdd522b5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622250 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a005177-344e-461b-b8ac-989f18ade02d-config\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622276 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32e4b213-a66b-4e97-83b1-cac96a741b9a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-64r5t\" (UID: \"32e4b213-a66b-4e97-83b1-cac96a741b9a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622302 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-config\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622329 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf423681-0ee6-4d44-bbed-e1447eaa7edd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-k98jk\" (UID: \"cf423681-0ee6-4d44-bbed-e1447eaa7edd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622359 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-config\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622384 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f8268de-4dd1-4521-9e5a-b4679996b511-etcd-client\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622406 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7e7175b7-4b92-49c8-98d1-b2366c099ced-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622425 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/50acf6a8-8d00-4bd2-9a88-f85cd8b4b504-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nnnkc\" (UID: \"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622445 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/90dd52f0-03cd-46fd-90cd-1a480bae1280-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xxvbh\" (UID: \"90dd52f0-03cd-46fd-90cd-1a480bae1280\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622466 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-serving-cert\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622486 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-client-ca\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622325 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8498b530-52d4-4ca5-88ca-fda75dfbb18a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622630 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7d71d251-484c-4198-b697-298f2d92031f-auth-proxy-config\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622674 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a005177-344e-461b-b8ac-989f18ade02d-trusted-ca\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622721 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622747 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622770 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7e7175b7-4b92-49c8-98d1-b2366c099ced-audit-dir\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622794 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdcq2\" (UniqueName: \"kubernetes.io/projected/cf423681-0ee6-4d44-bbed-e1447eaa7edd-kube-api-access-zdcq2\") pod \"openshift-apiserver-operator-796bbdcf4f-k98jk\" (UID: \"cf423681-0ee6-4d44-bbed-e1447eaa7edd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622818 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32e4b213-a66b-4e97-83b1-cac96a741b9a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-64r5t\" (UID: \"32e4b213-a66b-4e97-83b1-cac96a741b9a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622843 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd07f1f3-ea79-4612-af59-91151c9bb06f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vm2rn\" (UID: \"cd07f1f3-ea79-4612-af59-91151c9bb06f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622867 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9027bafb-1a39-4a53-9bcf-778b8859c192-config\") pod \"kube-apiserver-operator-766d6c64bb-58dfj\" (UID: \"9027bafb-1a39-4a53-9bcf-778b8859c192\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622890 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622900 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7e7175b7-4b92-49c8-98d1-b2366c099ced-audit-policies\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.622914 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-console-config\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.623157 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-sc29d"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.623606 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-console-config\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.623699 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.623911 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.624301 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf423681-0ee6-4d44-bbed-e1447eaa7edd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-k98jk\" (UID: \"cf423681-0ee6-4d44-bbed-e1447eaa7edd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.624393 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-config\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.624566 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-etcd-serving-ca\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.624868 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-oauth-serving-cert\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.624968 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.625110 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7e7175b7-4b92-49c8-98d1-b2366c099ced-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.625312 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.625557 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7d71d251-484c-4198-b697-298f2d92031f-auth-proxy-config\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.625673 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.625778 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dgtvg"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.625881 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7e7175b7-4b92-49c8-98d1-b2366c099ced-audit-dir\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.621772 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-service-ca\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.626905 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6376c9a4-9209-4f0d-835d-7db09de15649-available-featuregates\") pod \"openshift-config-operator-7777fb866f-m97k8\" (UID: \"6376c9a4-9209-4f0d-835d-7db09de15649\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.627122 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dgtvg"] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.627239 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.627346 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8498b530-52d4-4ca5-88ca-fda75dfbb18a-serving-cert\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.627569 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.627617 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/151dc193-527a-4506-bcb9-170641b9ea3f-audit-dir\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.639884 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-trusted-ca-bundle\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.639958 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/151dc193-527a-4506-bcb9-170641b9ea3f-audit\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.641598 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8498b530-52d4-4ca5-88ca-fda75dfbb18a-service-ca-bundle\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.642080 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7175b7-4b92-49c8-98d1-b2366c099ced-serving-cert\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.642846 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-serving-cert\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.643106 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8498b530-52d4-4ca5-88ca-fda75dfbb18a-config\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.643140 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0f7a4f62-f974-487f-90d0-f45d16638c4c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p8jqw\" (UID: \"0f7a4f62-f974-487f-90d0-f45d16638c4c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.644192 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6376c9a4-9209-4f0d-835d-7db09de15649-serving-cert\") pod \"openshift-config-operator-7777fb866f-m97k8\" (UID: \"6376c9a4-9209-4f0d-835d-7db09de15649\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.645118 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/151dc193-527a-4506-bcb9-170641b9ea3f-serving-cert\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.646109 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-oauth-config\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.646309 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.646956 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/151dc193-527a-4506-bcb9-170641b9ea3f-node-pullsecrets\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.647096 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7e7175b7-4b92-49c8-98d1-b2366c099ced-etcd-client\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.647104 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.647435 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7d71d251-484c-4198-b697-298f2d92031f-machine-approver-tls\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.641664 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7e7175b7-4b92-49c8-98d1-b2366c099ced-encryption-config\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.647736 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/151dc193-527a-4506-bcb9-170641b9ea3f-encryption-config\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.647842 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.647975 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/151dc193-527a-4506-bcb9-170641b9ea3f-node-pullsecrets\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.648092 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.649219 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.650460 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/151dc193-527a-4506-bcb9-170641b9ea3f-etcd-client\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.650773 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.651113 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7d71d251-484c-4198-b697-298f2d92031f-machine-approver-tls\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.651122 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf423681-0ee6-4d44-bbed-e1447eaa7edd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-k98jk\" (UID: \"cf423681-0ee6-4d44-bbed-e1447eaa7edd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.654421 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/151dc193-527a-4506-bcb9-170641b9ea3f-encryption-config\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.657835 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7e7175b7-4b92-49c8-98d1-b2366c099ced-etcd-client\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.668954 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.687348 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.706887 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.726935 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753070 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753380 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-787z7\" (UniqueName: \"kubernetes.io/projected/1cd73a43-1689-479b-bae3-6519bdd522b5-kube-api-access-787z7\") pod \"migrator-59844c95c7-42wg6\" (UID: \"1cd73a43-1689-479b-bae3-6519bdd522b5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753415 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a005177-344e-461b-b8ac-989f18ade02d-config\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753443 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32e4b213-a66b-4e97-83b1-cac96a741b9a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-64r5t\" (UID: \"32e4b213-a66b-4e97-83b1-cac96a741b9a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753471 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-config\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753502 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f8268de-4dd1-4521-9e5a-b4679996b511-etcd-client\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753528 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/50acf6a8-8d00-4bd2-9a88-f85cd8b4b504-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nnnkc\" (UID: \"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753572 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/90dd52f0-03cd-46fd-90cd-1a480bae1280-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xxvbh\" (UID: \"90dd52f0-03cd-46fd-90cd-1a480bae1280\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753595 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-client-ca\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753618 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a005177-344e-461b-b8ac-989f18ade02d-trusted-ca\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753655 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32e4b213-a66b-4e97-83b1-cac96a741b9a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-64r5t\" (UID: \"32e4b213-a66b-4e97-83b1-cac96a741b9a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753685 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd07f1f3-ea79-4612-af59-91151c9bb06f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vm2rn\" (UID: \"cd07f1f3-ea79-4612-af59-91151c9bb06f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753711 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9027bafb-1a39-4a53-9bcf-778b8859c192-config\") pod \"kube-apiserver-operator-766d6c64bb-58dfj\" (UID: \"9027bafb-1a39-4a53-9bcf-778b8859c192\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753740 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753763 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgp2g\" (UniqueName: \"kubernetes.io/projected/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-kube-api-access-zgp2g\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753790 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c6hh\" (UniqueName: \"kubernetes.io/projected/ab39813c-c2d7-432c-aa05-7b33c1e86fac-kube-api-access-5c6hh\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753818 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/242b082c-0e13-4b74-830e-0bb1008b5c0a-profile-collector-cert\") pod \"catalog-operator-68c6474976-ccjqw\" (UID: \"242b082c-0e13-4b74-830e-0bb1008b5c0a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753850 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f8268de-4dd1-4521-9e5a-b4679996b511-serving-cert\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753873 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd7j2\" (UniqueName: \"kubernetes.io/projected/32e4b213-a66b-4e97-83b1-cac96a741b9a-kube-api-access-pd7j2\") pod \"openshift-controller-manager-operator-756b6f6bc6-64r5t\" (UID: \"32e4b213-a66b-4e97-83b1-cac96a741b9a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753908 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/967b1700-c689-4195-8dcd-4177b702d066-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r4w72\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753948 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mgvd\" (UniqueName: \"kubernetes.io/projected/cd07f1f3-ea79-4612-af59-91151c9bb06f-kube-api-access-6mgvd\") pod \"control-plane-machine-set-operator-78cbb6b69f-vm2rn\" (UID: \"cd07f1f3-ea79-4612-af59-91151c9bb06f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753973 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/967b1700-c689-4195-8dcd-4177b702d066-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r4w72\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.753999 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r5jp\" (UniqueName: \"kubernetes.io/projected/fb1e412b-f423-4178-a306-b9201f089099-kube-api-access-5r5jp\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754050 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9027bafb-1a39-4a53-9bcf-778b8859c192-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-58dfj\" (UID: \"9027bafb-1a39-4a53-9bcf-778b8859c192\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754135 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab39813c-c2d7-432c-aa05-7b33c1e86fac-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754162 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b6ffb849-1700-439d-b7d3-1c1540a48fc1-webhook-cert\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754197 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9027bafb-1a39-4a53-9bcf-778b8859c192-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-58dfj\" (UID: \"9027bafb-1a39-4a53-9bcf-778b8859c192\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754229 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab39813c-c2d7-432c-aa05-7b33c1e86fac-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754254 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-serving-cert\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754284 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ab39813c-c2d7-432c-aa05-7b33c1e86fac-ready\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754313 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754339 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b6ffb849-1700-439d-b7d3-1c1540a48fc1-apiservice-cert\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754367 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754399 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx2rm\" (UniqueName: \"kubernetes.io/projected/4f8268de-4dd1-4521-9e5a-b4679996b511-kube-api-access-jx2rm\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754424 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb1e412b-f423-4178-a306-b9201f089099-images\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754448 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90dd52f0-03cd-46fd-90cd-1a480bae1280-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xxvbh\" (UID: \"90dd52f0-03cd-46fd-90cd-1a480bae1280\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754474 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/676cb1dc-3786-44f8-bf27-963da433b3c1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xlnmq\" (UID: \"676cb1dc-3786-44f8-bf27-963da433b3c1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754495 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b6ffb849-1700-439d-b7d3-1c1540a48fc1-tmpfs\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754519 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fb1e412b-f423-4178-a306-b9201f089099-proxy-tls\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754565 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjjfj\" (UniqueName: \"kubernetes.io/projected/967b1700-c689-4195-8dcd-4177b702d066-kube-api-access-mjjfj\") pod \"marketplace-operator-79b997595-r4w72\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754589 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754615 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fb1e412b-f423-4178-a306-b9201f089099-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754639 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtrk8\" (UniqueName: \"kubernetes.io/projected/242b082c-0e13-4b74-830e-0bb1008b5c0a-kube-api-access-gtrk8\") pod \"catalog-operator-68c6474976-ccjqw\" (UID: \"242b082c-0e13-4b74-830e-0bb1008b5c0a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754662 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfx84\" (UniqueName: \"kubernetes.io/projected/3a005177-344e-461b-b8ac-989f18ade02d-kube-api-access-bfx84\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754684 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4f8268de-4dd1-4521-9e5a-b4679996b511-etcd-ca\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754711 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f8268de-4dd1-4521-9e5a-b4679996b511-etcd-service-ca\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754739 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm2jn\" (UniqueName: \"kubernetes.io/projected/b6ffb849-1700-439d-b7d3-1c1540a48fc1-kube-api-access-fm2jn\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754763 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a005177-344e-461b-b8ac-989f18ade02d-serving-cert\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754786 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f8268de-4dd1-4521-9e5a-b4679996b511-config\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754808 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/50acf6a8-8d00-4bd2-9a88-f85cd8b4b504-proxy-tls\") pod \"machine-config-controller-84d6567774-nnnkc\" (UID: \"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754834 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w6w4\" (UniqueName: \"kubernetes.io/projected/50acf6a8-8d00-4bd2-9a88-f85cd8b4b504-kube-api-access-4w6w4\") pod \"machine-config-controller-84d6567774-nnnkc\" (UID: \"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754857 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90dd52f0-03cd-46fd-90cd-1a480bae1280-config\") pod \"kube-controller-manager-operator-78b949d7b-xxvbh\" (UID: \"90dd52f0-03cd-46fd-90cd-1a480bae1280\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754884 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7bnp\" (UniqueName: \"kubernetes.io/projected/676cb1dc-3786-44f8-bf27-963da433b3c1-kube-api-access-p7bnp\") pod \"multus-admission-controller-857f4d67dd-xlnmq\" (UID: \"676cb1dc-3786-44f8-bf27-963da433b3c1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.754909 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/242b082c-0e13-4b74-830e-0bb1008b5c0a-srv-cert\") pod \"catalog-operator-68c6474976-ccjqw\" (UID: \"242b082c-0e13-4b74-830e-0bb1008b5c0a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.755145 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:57.755122143 +0000 UTC m=+128.701489968 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.756040 4936 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.756212 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:57.756148451 +0000 UTC m=+128.702516276 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.757225 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fb1e412b-f423-4178-a306-b9201f089099-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.757248 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a005177-344e-461b-b8ac-989f18ade02d-config\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.758150 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4f8268de-4dd1-4521-9e5a-b4679996b511-etcd-ca\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.758652 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f8268de-4dd1-4521-9e5a-b4679996b511-config\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.758786 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32e4b213-a66b-4e97-83b1-cac96a741b9a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-64r5t\" (UID: \"32e4b213-a66b-4e97-83b1-cac96a741b9a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.758929 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f8268de-4dd1-4521-9e5a-b4679996b511-etcd-service-ca\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759122 4936 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759257 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759285 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759301 4936 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759261 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:57.759238505 +0000 UTC m=+128.705606320 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759469 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:57.759457631 +0000 UTC m=+128.705825446 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759156 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759485 4936 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759493 4936 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:41 crc kubenswrapper[4936]: E0320 16:02:41.759518 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 16:02:57.759510103 +0000 UTC m=+128.705877918 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.759638 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32e4b213-a66b-4e97-83b1-cac96a741b9a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-64r5t\" (UID: \"32e4b213-a66b-4e97-83b1-cac96a741b9a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.760099 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b6ffb849-1700-439d-b7d3-1c1540a48fc1-tmpfs\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.760406 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/50acf6a8-8d00-4bd2-9a88-f85cd8b4b504-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nnnkc\" (UID: \"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.760694 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a005177-344e-461b-b8ac-989f18ade02d-trusted-ca\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.760780 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ab39813c-c2d7-432c-aa05-7b33c1e86fac-ready\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.760980 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab39813c-c2d7-432c-aa05-7b33c1e86fac-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.761493 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-serving-cert\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.763162 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-config\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.767431 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-client-ca\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.769071 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a005177-344e-461b-b8ac-989f18ade02d-serving-cert\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.771835 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f8268de-4dd1-4521-9e5a-b4679996b511-serving-cert\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.774883 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtxgl\" (UniqueName: \"kubernetes.io/projected/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-kube-api-access-qtxgl\") pod \"controller-manager-879f6c89f-4p8zd\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.776930 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f8268de-4dd1-4521-9e5a-b4679996b511-etcd-client\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.786818 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.787403 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wstk\" (UniqueName: \"kubernetes.io/projected/d815687f-6e2a-44c8-bf71-c9718db7ba67-kube-api-access-6wstk\") pod \"machine-api-operator-5694c8668f-ssr5n\" (UID: \"d815687f-6e2a-44c8-bf71-c9718db7ba67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.827652 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.847262 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.853332 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.853332 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.853506 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.866836 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.870373 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb1e412b-f423-4178-a306-b9201f089099-images\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.885951 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.909990 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.926378 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.932579 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b6ffb849-1700-439d-b7d3-1c1540a48fc1-apiservice-cert\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.935074 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b6ffb849-1700-439d-b7d3-1c1540a48fc1-webhook-cert\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.947240 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.954634 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.955344 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fb1e412b-f423-4178-a306-b9201f089099-proxy-tls\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.967010 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.969007 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" Mar 20 16:02:41 crc kubenswrapper[4936]: I0320 16:02:41.987207 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.007259 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.016351 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd07f1f3-ea79-4612-af59-91151c9bb06f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vm2rn\" (UID: \"cd07f1f3-ea79-4612-af59-91151c9bb06f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.029066 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.050022 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.057834 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90dd52f0-03cd-46fd-90cd-1a480bae1280-config\") pod \"kube-controller-manager-operator-78b949d7b-xxvbh\" (UID: \"90dd52f0-03cd-46fd-90cd-1a480bae1280\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.066948 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.089539 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.112268 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.127514 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.147391 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.154810 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90dd52f0-03cd-46fd-90cd-1a480bae1280-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xxvbh\" (UID: \"90dd52f0-03cd-46fd-90cd-1a480bae1280\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.171479 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.172859 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4p8zd"] Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.179924 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/242b082c-0e13-4b74-830e-0bb1008b5c0a-srv-cert\") pod \"catalog-operator-68c6474976-ccjqw\" (UID: \"242b082c-0e13-4b74-830e-0bb1008b5c0a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.191235 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.195617 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ssr5n"] Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.202753 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/242b082c-0e13-4b74-830e-0bb1008b5c0a-profile-collector-cert\") pod \"catalog-operator-68c6474976-ccjqw\" (UID: \"242b082c-0e13-4b74-830e-0bb1008b5c0a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:42 crc kubenswrapper[4936]: W0320 16:02:42.204824 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd815687f_6e2a_44c8_bf71_c9718db7ba67.slice/crio-e953cb62d68c10f2503fe78dc4c87e6571c3fa9464852f43b6997f3c4df866dc WatchSource:0}: Error finding container e953cb62d68c10f2503fe78dc4c87e6571c3fa9464852f43b6997f3c4df866dc: Status 404 returned error can't find the container with id e953cb62d68c10f2503fe78dc4c87e6571c3fa9464852f43b6997f3c4df866dc Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.208365 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.222401 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/50acf6a8-8d00-4bd2-9a88-f85cd8b4b504-proxy-tls\") pod \"machine-config-controller-84d6567774-nnnkc\" (UID: \"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.230449 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.247603 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.266691 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.273104 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/676cb1dc-3786-44f8-bf27-963da433b3c1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xlnmq\" (UID: \"676cb1dc-3786-44f8-bf27-963da433b3c1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.287492 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.296837 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab39813c-c2d7-432c-aa05-7b33c1e86fac-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.313395 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.323758 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/967b1700-c689-4195-8dcd-4177b702d066-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r4w72\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.327013 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.336408 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/967b1700-c689-4195-8dcd-4177b702d066-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r4w72\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.348312 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.367513 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.389496 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.408646 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.426887 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.435928 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9027bafb-1a39-4a53-9bcf-778b8859c192-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-58dfj\" (UID: \"9027bafb-1a39-4a53-9bcf-778b8859c192\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.446745 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.450448 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9027bafb-1a39-4a53-9bcf-778b8859c192-config\") pod \"kube-apiserver-operator-766d6c64bb-58dfj\" (UID: \"9027bafb-1a39-4a53-9bcf-778b8859c192\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.465787 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.507773 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.519513 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" event={"ID":"d815687f-6e2a-44c8-bf71-c9718db7ba67","Type":"ContainerStarted","Data":"899ffc018c091f1a5ef53c62c4f053c212a180d4868033971e28fd99d9c014f1"} Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.519592 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" event={"ID":"d815687f-6e2a-44c8-bf71-c9718db7ba67","Type":"ContainerStarted","Data":"f8761094110726ede89996d399ab1818434125a782a56a4be44e853655bf759a"} Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.519609 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" event={"ID":"d815687f-6e2a-44c8-bf71-c9718db7ba67","Type":"ContainerStarted","Data":"e953cb62d68c10f2503fe78dc4c87e6571c3fa9464852f43b6997f3c4df866dc"} Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.521064 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" event={"ID":"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21","Type":"ContainerStarted","Data":"bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca"} Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.521082 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" event={"ID":"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21","Type":"ContainerStarted","Data":"91b9503fdd4ee9b28c1eae0da7d03d7a07f8ba84c60c9cc52a859fbbb820bc82"} Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.521693 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.527411 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.528809 4936 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4p8zd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.528864 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" podUID="20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.544970 4936 request.go:700] Waited for 1.009698522s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-scheduler-operator/secrets?fieldSelector=metadata.name%3Dkube-scheduler-operator-serving-cert&limit=500&resourceVersion=0 Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.546842 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.567502 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.595532 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.606819 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.627240 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.646692 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.668194 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.686294 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.707403 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.726891 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.747924 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.767800 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.788056 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.806936 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.827369 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.847491 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.853534 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.867356 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.888096 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.908014 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.928216 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.947868 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.969317 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 20 16:02:42 crc kubenswrapper[4936]: I0320 16:02:42.988235 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.006808 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.027831 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.047511 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.067850 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.086768 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.107870 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.128229 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.146431 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.166478 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.187879 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.207683 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.228010 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.247192 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.266698 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.287257 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.307663 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.326885 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.376512 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwssr\" (UniqueName: \"kubernetes.io/projected/1f1994bb-a56f-4e25-b6d8-7bcc2113de10-kube-api-access-bwssr\") pod \"dns-operator-744455d44c-q6cx6\" (UID: \"1f1994bb-a56f-4e25-b6d8-7bcc2113de10\") " pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.388124 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.389044 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5llft\" (UniqueName: \"kubernetes.io/projected/7d71d251-484c-4198-b697-298f2d92031f-kube-api-access-5llft\") pod \"machine-approver-56656f9798-vqmvx\" (UID: \"7d71d251-484c-4198-b697-298f2d92031f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.406673 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.427968 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.483430 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdcq2\" (UniqueName: \"kubernetes.io/projected/cf423681-0ee6-4d44-bbed-e1447eaa7edd-kube-api-access-zdcq2\") pod \"openshift-apiserver-operator-796bbdcf4f-k98jk\" (UID: \"cf423681-0ee6-4d44-bbed-e1447eaa7edd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.483788 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.486331 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.496279 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg5hx\" (UniqueName: \"kubernetes.io/projected/9333cf9d-b80d-4242-a940-c92ed3b593d1-kube-api-access-rg5hx\") pod \"oauth-openshift-558db77b4-jhqvm\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.507062 4936 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.511653 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.529899 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.535152 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.545867 4936 request.go:700] Waited for 1.906483635s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-config-operator/serviceaccounts/openshift-config-operator/token Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.569055 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4gzw\" (UniqueName: \"kubernetes.io/projected/6376c9a4-9209-4f0d-835d-7db09de15649-kube-api-access-j4gzw\") pod \"openshift-config-operator-7777fb866f-m97k8\" (UID: \"6376c9a4-9209-4f0d-835d-7db09de15649\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.588945 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.592298 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc7q8\" (UniqueName: \"kubernetes.io/projected/151dc193-527a-4506-bcb9-170641b9ea3f-kube-api-access-zc7q8\") pod \"apiserver-76f77b778f-jgfjb\" (UID: \"151dc193-527a-4506-bcb9-170641b9ea3f\") " pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.614082 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frvx8\" (UniqueName: \"kubernetes.io/projected/bf0f9e97-bddb-4054-99dd-d10428233e9f-kube-api-access-frvx8\") pod \"downloads-7954f5f757-dzgl2\" (UID: \"bf0f9e97-bddb-4054-99dd-d10428233e9f\") " pod="openshift-console/downloads-7954f5f757-dzgl2" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.617286 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dzgl2" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.623214 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48jmz\" (UniqueName: \"kubernetes.io/projected/8498b530-52d4-4ca5-88ca-fda75dfbb18a-kube-api-access-48jmz\") pod \"authentication-operator-69f744f599-dkllp\" (UID: \"8498b530-52d4-4ca5-88ca-fda75dfbb18a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.646350 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpxkx\" (UniqueName: \"kubernetes.io/projected/c52c0329-ca68-4d95-8341-ab8191c3a186-kube-api-access-wpxkx\") pod \"console-f9d7485db-6q5fl\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.647702 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.655878 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.665785 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj2cd\" (UniqueName: \"kubernetes.io/projected/7e7175b7-4b92-49c8-98d1-b2366c099ced-kube-api-access-nj2cd\") pod \"apiserver-7bbb656c7d-57jx7\" (UID: \"7e7175b7-4b92-49c8-98d1-b2366c099ced\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.699803 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5ghv\" (UniqueName: \"kubernetes.io/projected/0f7a4f62-f974-487f-90d0-f45d16638c4c-kube-api-access-f5ghv\") pod \"cluster-samples-operator-665b6dd947-p8jqw\" (UID: \"0f7a4f62-f974-487f-90d0-f45d16638c4c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.705166 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-787z7\" (UniqueName: \"kubernetes.io/projected/1cd73a43-1689-479b-bae3-6519bdd522b5-kube-api-access-787z7\") pod \"migrator-59844c95c7-42wg6\" (UID: \"1cd73a43-1689-479b-bae3-6519bdd522b5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.725333 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm2jn\" (UniqueName: \"kubernetes.io/projected/b6ffb849-1700-439d-b7d3-1c1540a48fc1-kube-api-access-fm2jn\") pod \"packageserver-d55dfcdfc-2w76v\" (UID: \"b6ffb849-1700-439d-b7d3-1c1540a48fc1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.742190 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk"] Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.769457 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.794533 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtrk8\" (UniqueName: \"kubernetes.io/projected/242b082c-0e13-4b74-830e-0bb1008b5c0a-kube-api-access-gtrk8\") pod \"catalog-operator-68c6474976-ccjqw\" (UID: \"242b082c-0e13-4b74-830e-0bb1008b5c0a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.799586 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jhqvm"] Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.800210 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.802245 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.803037 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfx84\" (UniqueName: \"kubernetes.io/projected/3a005177-344e-461b-b8ac-989f18ade02d-kube-api-access-bfx84\") pod \"console-operator-58897d9998-xtqwz\" (UID: \"3a005177-344e-461b-b8ac-989f18ade02d\") " pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.817359 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w6w4\" (UniqueName: \"kubernetes.io/projected/50acf6a8-8d00-4bd2-9a88-f85cd8b4b504-kube-api-access-4w6w4\") pod \"machine-config-controller-84d6567774-nnnkc\" (UID: \"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.817721 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.817956 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.824382 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c6hh\" (UniqueName: \"kubernetes.io/projected/ab39813c-c2d7-432c-aa05-7b33c1e86fac-kube-api-access-5c6hh\") pod \"cni-sysctl-allowlist-ds-m5rqk\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.833398 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.834685 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.839362 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.868102 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/90dd52f0-03cd-46fd-90cd-1a480bae1280-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xxvbh\" (UID: \"90dd52f0-03cd-46fd-90cd-1a480bae1280\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.871704 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.875807 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx2rm\" (UniqueName: \"kubernetes.io/projected/4f8268de-4dd1-4521-9e5a-b4679996b511-kube-api-access-jx2rm\") pod \"etcd-operator-b45778765-j54jw\" (UID: \"4f8268de-4dd1-4521-9e5a-b4679996b511\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.877076 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgp2g\" (UniqueName: \"kubernetes.io/projected/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-kube-api-access-zgp2g\") pod \"route-controller-manager-6576b87f9c-cf7gl\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.928588 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mgvd\" (UniqueName: \"kubernetes.io/projected/cd07f1f3-ea79-4612-af59-91151c9bb06f-kube-api-access-6mgvd\") pod \"control-plane-machine-set-operator-78cbb6b69f-vm2rn\" (UID: \"cd07f1f3-ea79-4612-af59-91151c9bb06f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.935251 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9027bafb-1a39-4a53-9bcf-778b8859c192-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-58dfj\" (UID: \"9027bafb-1a39-4a53-9bcf-778b8859c192\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.935596 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjjfj\" (UniqueName: \"kubernetes.io/projected/967b1700-c689-4195-8dcd-4177b702d066-kube-api-access-mjjfj\") pod \"marketplace-operator-79b997595-r4w72\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.949178 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.966712 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7bnp\" (UniqueName: \"kubernetes.io/projected/676cb1dc-3786-44f8-bf27-963da433b3c1-kube-api-access-p7bnp\") pod \"multus-admission-controller-857f4d67dd-xlnmq\" (UID: \"676cb1dc-3786-44f8-bf27-963da433b3c1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.969847 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.987404 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.995727 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r5jp\" (UniqueName: \"kubernetes.io/projected/fb1e412b-f423-4178-a306-b9201f089099-kube-api-access-5r5jp\") pod \"machine-config-operator-74547568cd-qxmr7\" (UID: \"fb1e412b-f423-4178-a306-b9201f089099\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.996259 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd7j2\" (UniqueName: \"kubernetes.io/projected/32e4b213-a66b-4e97-83b1-cac96a741b9a-kube-api-access-pd7j2\") pod \"openshift-controller-manager-operator-756b6f6bc6-64r5t\" (UID: \"32e4b213-a66b-4e97-83b1-cac96a741b9a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:43 crc kubenswrapper[4936]: I0320 16:02:43.999454 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.007648 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.021887 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.027955 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.035868 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dzgl2"] Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.049002 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.066799 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.095828 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.098464 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.098629 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.106604 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.124998 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.125454 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.152885 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.177250 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.205960 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-trusted-ca\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206011 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53d1ed6e-605c-46b9-b053-ba5a899db885-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206057 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvz5h\" (UniqueName: \"kubernetes.io/projected/5d3858e7-c19f-4140-847c-7742c0af0d9c-kube-api-access-mvz5h\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206107 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d3858e7-c19f-4140-847c-7742c0af0d9c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206129 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d3858e7-c19f-4140-847c-7742c0af0d9c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206166 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g45pg\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-kube-api-access-g45pg\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206197 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d3858e7-c19f-4140-847c-7742c0af0d9c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206220 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-certificates\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206255 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206284 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-bound-sa-token\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206312 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-tls\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.206568 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53d1ed6e-605c-46b9-b053-ba5a899db885-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: E0320 16:02:44.206913 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:44.706882025 +0000 UTC m=+115.653249840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.244084 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q6cx6"] Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.310837 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311214 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53d1ed6e-605c-46b9-b053-ba5a899db885-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311275 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvz5h\" (UniqueName: \"kubernetes.io/projected/5d3858e7-c19f-4140-847c-7742c0af0d9c-kube-api-access-mvz5h\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311298 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ed539e-3d39-44e4-a7c9-15f51fae30ae-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tvs2g\" (UID: \"95ed539e-3d39-44e4-a7c9-15f51fae30ae\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311316 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0a7511a-0072-4933-bd7f-42f1203900ae-profile-collector-cert\") pod \"olm-operator-6b444d44fb-l96sz\" (UID: \"b0a7511a-0072-4933-bd7f-42f1203900ae\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311396 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d3858e7-c19f-4140-847c-7742c0af0d9c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311427 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqq8l\" (UniqueName: \"kubernetes.io/projected/921365e5-3846-485e-9afa-270834e81db5-kube-api-access-bqq8l\") pod \"dns-default-5nbl4\" (UID: \"921365e5-3846-485e-9afa-270834e81db5\") " pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311449 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8j5b\" (UniqueName: \"kubernetes.io/projected/74a1703c-e83e-487e-9d98-f2d948e3f831-kube-api-access-h8j5b\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311467 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62xb4\" (UniqueName: \"kubernetes.io/projected/6b4d3e08-82bf-4642-8145-cb6b1db82ad6-kube-api-access-62xb4\") pod \"machine-config-server-sc29d\" (UID: \"6b4d3e08-82bf-4642-8145-cb6b1db82ad6\") " pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311486 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d3858e7-c19f-4140-847c-7742c0af0d9c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311502 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/74a1703c-e83e-487e-9d98-f2d948e3f831-metrics-tls\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311536 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1766c31e-3eb1-4723-9fa9-4768f826f8ec-service-ca-bundle\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311631 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gxdw\" (UniqueName: \"kubernetes.io/projected/e742a861-00e7-4ddc-b9d9-ab5dd22c84e2-kube-api-access-6gxdw\") pod \"service-ca-9c57cc56f-8v7wv\" (UID: \"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311660 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g45pg\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-kube-api-access-g45pg\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311707 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhp77\" (UniqueName: \"kubernetes.io/projected/e84e4377-e478-4c12-85bf-b236885152b2-kube-api-access-rhp77\") pod \"collect-profiles-29567040-9g8q5\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311729 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnvgp\" (UniqueName: \"kubernetes.io/projected/1766c31e-3eb1-4723-9fa9-4768f826f8ec-kube-api-access-vnvgp\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311761 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-socket-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311779 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c11e87a-9743-40b0-8741-dfec0ce96cb3-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wspgg\" (UID: \"2c11e87a-9743-40b0-8741-dfec0ce96cb3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311823 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95ed539e-3d39-44e4-a7c9-15f51fae30ae-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tvs2g\" (UID: \"95ed539e-3d39-44e4-a7c9-15f51fae30ae\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311840 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68cjb\" (UniqueName: \"kubernetes.io/projected/2d469c13-4b04-4c81-9b1b-177250681db4-kube-api-access-68cjb\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311901 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-mountpoint-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311929 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d3858e7-c19f-4140-847c-7742c0af0d9c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311974 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-certificates\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.311989 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1766c31e-3eb1-4723-9fa9-4768f826f8ec-default-certificate\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312018 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5l6qv\" (UID: \"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312088 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0a7511a-0072-4933-bd7f-42f1203900ae-srv-cert\") pod \"olm-operator-6b444d44fb-l96sz\" (UID: \"b0a7511a-0072-4933-bd7f-42f1203900ae\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312127 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-bound-sa-token\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312144 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5l6qv\" (UID: \"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312224 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-tls\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312248 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-748h8\" (UniqueName: \"kubernetes.io/projected/aad2185e-c38e-4b5e-9253-5499237e913c-kube-api-access-748h8\") pod \"service-ca-operator-777779d784-nsh5l\" (UID: \"aad2185e-c38e-4b5e-9253-5499237e913c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312321 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e742a861-00e7-4ddc-b9d9-ab5dd22c84e2-signing-cabundle\") pod \"service-ca-9c57cc56f-8v7wv\" (UID: \"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312357 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1766c31e-3eb1-4723-9fa9-4768f826f8ec-metrics-certs\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312389 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/921365e5-3846-485e-9afa-270834e81db5-config-volume\") pod \"dns-default-5nbl4\" (UID: \"921365e5-3846-485e-9afa-270834e81db5\") " pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312408 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kxn4\" (UniqueName: \"kubernetes.io/projected/b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4-kube-api-access-6kxn4\") pod \"kube-storage-version-migrator-operator-b67b599dd-5l6qv\" (UID: \"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312427 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8v6h\" (UniqueName: \"kubernetes.io/projected/b0a7511a-0072-4933-bd7f-42f1203900ae-kube-api-access-j8v6h\") pod \"olm-operator-6b444d44fb-l96sz\" (UID: \"b0a7511a-0072-4933-bd7f-42f1203900ae\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312451 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/921365e5-3846-485e-9afa-270834e81db5-metrics-tls\") pod \"dns-default-5nbl4\" (UID: \"921365e5-3846-485e-9afa-270834e81db5\") " pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312466 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95ed539e-3d39-44e4-a7c9-15f51fae30ae-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tvs2g\" (UID: \"95ed539e-3d39-44e4-a7c9-15f51fae30ae\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312481 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74a1703c-e83e-487e-9d98-f2d948e3f831-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312530 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84e4377-e478-4c12-85bf-b236885152b2-secret-volume\") pod \"collect-profiles-29567040-9g8q5\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312562 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-csi-data-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312583 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgcgg\" (UniqueName: \"kubernetes.io/projected/2c11e87a-9743-40b0-8741-dfec0ce96cb3-kube-api-access-jgcgg\") pod \"package-server-manager-789f6589d5-wspgg\" (UID: \"2c11e87a-9743-40b0-8741-dfec0ce96cb3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312600 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1766c31e-3eb1-4723-9fa9-4768f826f8ec-stats-auth\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312638 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6b4d3e08-82bf-4642-8145-cb6b1db82ad6-certs\") pod \"machine-config-server-sc29d\" (UID: \"6b4d3e08-82bf-4642-8145-cb6b1db82ad6\") " pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312713 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-plugins-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312751 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6b4d3e08-82bf-4642-8145-cb6b1db82ad6-node-bootstrap-token\") pod \"machine-config-server-sc29d\" (UID: \"6b4d3e08-82bf-4642-8145-cb6b1db82ad6\") " pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312780 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aad2185e-c38e-4b5e-9253-5499237e913c-config\") pod \"service-ca-operator-777779d784-nsh5l\" (UID: \"aad2185e-c38e-4b5e-9253-5499237e913c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312795 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e742a861-00e7-4ddc-b9d9-ab5dd22c84e2-signing-key\") pod \"service-ca-9c57cc56f-8v7wv\" (UID: \"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312823 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53d1ed6e-605c-46b9-b053-ba5a899db885-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312851 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aad2185e-c38e-4b5e-9253-5499237e913c-serving-cert\") pod \"service-ca-operator-777779d784-nsh5l\" (UID: \"aad2185e-c38e-4b5e-9253-5499237e913c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312868 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0abe164-30be-4265-b4b2-4ce8c7b24597-cert\") pod \"ingress-canary-qbm7q\" (UID: \"a0abe164-30be-4265-b4b2-4ce8c7b24597\") " pod="openshift-ingress-canary/ingress-canary-qbm7q" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312887 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-trusted-ca\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312912 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74a1703c-e83e-487e-9d98-f2d948e3f831-trusted-ca\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312931 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zprt6\" (UniqueName: \"kubernetes.io/projected/a0abe164-30be-4265-b4b2-4ce8c7b24597-kube-api-access-zprt6\") pod \"ingress-canary-qbm7q\" (UID: \"a0abe164-30be-4265-b4b2-4ce8c7b24597\") " pod="openshift-ingress-canary/ingress-canary-qbm7q" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312947 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84e4377-e478-4c12-85bf-b236885152b2-config-volume\") pod \"collect-profiles-29567040-9g8q5\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.312963 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-registration-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: E0320 16:02:44.313074 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:44.813039649 +0000 UTC m=+115.759407464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.315744 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d3858e7-c19f-4140-847c-7742c0af0d9c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.323786 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-certificates\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.326996 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53d1ed6e-605c-46b9-b053-ba5a899db885-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.328039 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-trusted-ca\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.343384 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-m97k8"] Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.357930 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d3858e7-c19f-4140-847c-7742c0af0d9c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.359643 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53d1ed6e-605c-46b9-b053-ba5a899db885-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.364570 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-tls\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.365951 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v"] Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.379731 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvz5h\" (UniqueName: \"kubernetes.io/projected/5d3858e7-c19f-4140-847c-7742c0af0d9c-kube-api-access-mvz5h\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.381845 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g45pg\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-kube-api-access-g45pg\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.388027 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-bound-sa-token\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.414089 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqq8l\" (UniqueName: \"kubernetes.io/projected/921365e5-3846-485e-9afa-270834e81db5-kube-api-access-bqq8l\") pod \"dns-default-5nbl4\" (UID: \"921365e5-3846-485e-9afa-270834e81db5\") " pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.414141 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8j5b\" (UniqueName: \"kubernetes.io/projected/74a1703c-e83e-487e-9d98-f2d948e3f831-kube-api-access-h8j5b\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.414172 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62xb4\" (UniqueName: \"kubernetes.io/projected/6b4d3e08-82bf-4642-8145-cb6b1db82ad6-kube-api-access-62xb4\") pod \"machine-config-server-sc29d\" (UID: \"6b4d3e08-82bf-4642-8145-cb6b1db82ad6\") " pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.414200 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/74a1703c-e83e-487e-9d98-f2d948e3f831-metrics-tls\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.414220 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1766c31e-3eb1-4723-9fa9-4768f826f8ec-service-ca-bundle\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.414250 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gxdw\" (UniqueName: \"kubernetes.io/projected/e742a861-00e7-4ddc-b9d9-ab5dd22c84e2-kube-api-access-6gxdw\") pod \"service-ca-9c57cc56f-8v7wv\" (UID: \"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.414268 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhp77\" (UniqueName: \"kubernetes.io/projected/e84e4377-e478-4c12-85bf-b236885152b2-kube-api-access-rhp77\") pod \"collect-profiles-29567040-9g8q5\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.414287 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnvgp\" (UniqueName: \"kubernetes.io/projected/1766c31e-3eb1-4723-9fa9-4768f826f8ec-kube-api-access-vnvgp\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.414304 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-socket-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416126 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c11e87a-9743-40b0-8741-dfec0ce96cb3-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wspgg\" (UID: \"2c11e87a-9743-40b0-8741-dfec0ce96cb3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416163 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95ed539e-3d39-44e4-a7c9-15f51fae30ae-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tvs2g\" (UID: \"95ed539e-3d39-44e4-a7c9-15f51fae30ae\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416192 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68cjb\" (UniqueName: \"kubernetes.io/projected/2d469c13-4b04-4c81-9b1b-177250681db4-kube-api-access-68cjb\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416222 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-mountpoint-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416261 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1766c31e-3eb1-4723-9fa9-4768f826f8ec-default-certificate\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416327 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5l6qv\" (UID: \"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416370 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416401 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0a7511a-0072-4933-bd7f-42f1203900ae-srv-cert\") pod \"olm-operator-6b444d44fb-l96sz\" (UID: \"b0a7511a-0072-4933-bd7f-42f1203900ae\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416438 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5l6qv\" (UID: \"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416483 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-748h8\" (UniqueName: \"kubernetes.io/projected/aad2185e-c38e-4b5e-9253-5499237e913c-kube-api-access-748h8\") pod \"service-ca-operator-777779d784-nsh5l\" (UID: \"aad2185e-c38e-4b5e-9253-5499237e913c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416522 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e742a861-00e7-4ddc-b9d9-ab5dd22c84e2-signing-cabundle\") pod \"service-ca-9c57cc56f-8v7wv\" (UID: \"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416597 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/921365e5-3846-485e-9afa-270834e81db5-config-volume\") pod \"dns-default-5nbl4\" (UID: \"921365e5-3846-485e-9afa-270834e81db5\") " pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416617 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kxn4\" (UniqueName: \"kubernetes.io/projected/b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4-kube-api-access-6kxn4\") pod \"kube-storage-version-migrator-operator-b67b599dd-5l6qv\" (UID: \"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416632 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1766c31e-3eb1-4723-9fa9-4768f826f8ec-metrics-certs\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416651 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/921365e5-3846-485e-9afa-270834e81db5-metrics-tls\") pod \"dns-default-5nbl4\" (UID: \"921365e5-3846-485e-9afa-270834e81db5\") " pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416668 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95ed539e-3d39-44e4-a7c9-15f51fae30ae-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tvs2g\" (UID: \"95ed539e-3d39-44e4-a7c9-15f51fae30ae\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416684 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8v6h\" (UniqueName: \"kubernetes.io/projected/b0a7511a-0072-4933-bd7f-42f1203900ae-kube-api-access-j8v6h\") pod \"olm-operator-6b444d44fb-l96sz\" (UID: \"b0a7511a-0072-4933-bd7f-42f1203900ae\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416713 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74a1703c-e83e-487e-9d98-f2d948e3f831-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416737 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84e4377-e478-4c12-85bf-b236885152b2-secret-volume\") pod \"collect-profiles-29567040-9g8q5\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416755 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-csi-data-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416774 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgcgg\" (UniqueName: \"kubernetes.io/projected/2c11e87a-9743-40b0-8741-dfec0ce96cb3-kube-api-access-jgcgg\") pod \"package-server-manager-789f6589d5-wspgg\" (UID: \"2c11e87a-9743-40b0-8741-dfec0ce96cb3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416790 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1766c31e-3eb1-4723-9fa9-4768f826f8ec-stats-auth\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416819 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6b4d3e08-82bf-4642-8145-cb6b1db82ad6-certs\") pod \"machine-config-server-sc29d\" (UID: \"6b4d3e08-82bf-4642-8145-cb6b1db82ad6\") " pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416845 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-plugins-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416864 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6b4d3e08-82bf-4642-8145-cb6b1db82ad6-node-bootstrap-token\") pod \"machine-config-server-sc29d\" (UID: \"6b4d3e08-82bf-4642-8145-cb6b1db82ad6\") " pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416883 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aad2185e-c38e-4b5e-9253-5499237e913c-config\") pod \"service-ca-operator-777779d784-nsh5l\" (UID: \"aad2185e-c38e-4b5e-9253-5499237e913c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416898 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e742a861-00e7-4ddc-b9d9-ab5dd22c84e2-signing-key\") pod \"service-ca-9c57cc56f-8v7wv\" (UID: \"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416919 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aad2185e-c38e-4b5e-9253-5499237e913c-serving-cert\") pod \"service-ca-operator-777779d784-nsh5l\" (UID: \"aad2185e-c38e-4b5e-9253-5499237e913c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416935 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0abe164-30be-4265-b4b2-4ce8c7b24597-cert\") pod \"ingress-canary-qbm7q\" (UID: \"a0abe164-30be-4265-b4b2-4ce8c7b24597\") " pod="openshift-ingress-canary/ingress-canary-qbm7q" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416958 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74a1703c-e83e-487e-9d98-f2d948e3f831-trusted-ca\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416975 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zprt6\" (UniqueName: \"kubernetes.io/projected/a0abe164-30be-4265-b4b2-4ce8c7b24597-kube-api-access-zprt6\") pod \"ingress-canary-qbm7q\" (UID: \"a0abe164-30be-4265-b4b2-4ce8c7b24597\") " pod="openshift-ingress-canary/ingress-canary-qbm7q" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.416995 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-registration-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.417013 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84e4377-e478-4c12-85bf-b236885152b2-config-volume\") pod \"collect-profiles-29567040-9g8q5\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.417059 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ed539e-3d39-44e4-a7c9-15f51fae30ae-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tvs2g\" (UID: \"95ed539e-3d39-44e4-a7c9-15f51fae30ae\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.417082 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0a7511a-0072-4933-bd7f-42f1203900ae-profile-collector-cert\") pod \"olm-operator-6b444d44fb-l96sz\" (UID: \"b0a7511a-0072-4933-bd7f-42f1203900ae\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: E0320 16:02:44.417161 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:44.917134597 +0000 UTC m=+115.863502612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.417359 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1766c31e-3eb1-4723-9fa9-4768f826f8ec-service-ca-bundle\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.417926 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-mountpoint-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.418127 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-socket-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.418577 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5l6qv\" (UID: \"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.423268 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-plugins-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.424267 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/921365e5-3846-485e-9afa-270834e81db5-config-volume\") pod \"dns-default-5nbl4\" (UID: \"921365e5-3846-485e-9afa-270834e81db5\") " pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.424824 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aad2185e-c38e-4b5e-9253-5499237e913c-config\") pod \"service-ca-operator-777779d784-nsh5l\" (UID: \"aad2185e-c38e-4b5e-9253-5499237e913c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.425994 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-csi-data-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.426752 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74a1703c-e83e-487e-9d98-f2d948e3f831-trusted-ca\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.427076 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ed539e-3d39-44e4-a7c9-15f51fae30ae-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tvs2g\" (UID: \"95ed539e-3d39-44e4-a7c9-15f51fae30ae\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.427876 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e742a861-00e7-4ddc-b9d9-ab5dd22c84e2-signing-cabundle\") pod \"service-ca-9c57cc56f-8v7wv\" (UID: \"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.428712 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2d469c13-4b04-4c81-9b1b-177250681db4-registration-dir\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.431336 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0a7511a-0072-4933-bd7f-42f1203900ae-srv-cert\") pod \"olm-operator-6b444d44fb-l96sz\" (UID: \"b0a7511a-0072-4933-bd7f-42f1203900ae\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.439343 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/74a1703c-e83e-487e-9d98-f2d948e3f831-metrics-tls\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.441296 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/921365e5-3846-485e-9afa-270834e81db5-metrics-tls\") pod \"dns-default-5nbl4\" (UID: \"921365e5-3846-485e-9afa-270834e81db5\") " pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.441495 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84e4377-e478-4c12-85bf-b236885152b2-config-volume\") pod \"collect-profiles-29567040-9g8q5\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.443805 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1766c31e-3eb1-4723-9fa9-4768f826f8ec-default-certificate\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.444103 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1766c31e-3eb1-4723-9fa9-4768f826f8ec-stats-auth\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.449584 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0a7511a-0072-4933-bd7f-42f1203900ae-profile-collector-cert\") pod \"olm-operator-6b444d44fb-l96sz\" (UID: \"b0a7511a-0072-4933-bd7f-42f1203900ae\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.466445 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e742a861-00e7-4ddc-b9d9-ab5dd22c84e2-signing-key\") pod \"service-ca-9c57cc56f-8v7wv\" (UID: \"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.466495 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0abe164-30be-4265-b4b2-4ce8c7b24597-cert\") pod \"ingress-canary-qbm7q\" (UID: \"a0abe164-30be-4265-b4b2-4ce8c7b24597\") " pod="openshift-ingress-canary/ingress-canary-qbm7q" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.466641 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d3858e7-c19f-4140-847c-7742c0af0d9c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-slzkt\" (UID: \"5d3858e7-c19f-4140-847c-7742c0af0d9c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.466956 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84e4377-e478-4c12-85bf-b236885152b2-secret-volume\") pod \"collect-profiles-29567040-9g8q5\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.466997 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1766c31e-3eb1-4723-9fa9-4768f826f8ec-metrics-certs\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.469136 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95ed539e-3d39-44e4-a7c9-15f51fae30ae-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tvs2g\" (UID: \"95ed539e-3d39-44e4-a7c9-15f51fae30ae\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.469251 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6b4d3e08-82bf-4642-8145-cb6b1db82ad6-certs\") pod \"machine-config-server-sc29d\" (UID: \"6b4d3e08-82bf-4642-8145-cb6b1db82ad6\") " pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.472191 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aad2185e-c38e-4b5e-9253-5499237e913c-serving-cert\") pod \"service-ca-operator-777779d784-nsh5l\" (UID: \"aad2185e-c38e-4b5e-9253-5499237e913c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.472205 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6b4d3e08-82bf-4642-8145-cb6b1db82ad6-node-bootstrap-token\") pod \"machine-config-server-sc29d\" (UID: \"6b4d3e08-82bf-4642-8145-cb6b1db82ad6\") " pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.472559 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5l6qv\" (UID: \"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.478175 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c11e87a-9743-40b0-8741-dfec0ce96cb3-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wspgg\" (UID: \"2c11e87a-9743-40b0-8741-dfec0ce96cb3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.485615 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8j5b\" (UniqueName: \"kubernetes.io/projected/74a1703c-e83e-487e-9d98-f2d948e3f831-kube-api-access-h8j5b\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.492057 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqq8l\" (UniqueName: \"kubernetes.io/projected/921365e5-3846-485e-9afa-270834e81db5-kube-api-access-bqq8l\") pod \"dns-default-5nbl4\" (UID: \"921365e5-3846-485e-9afa-270834e81db5\") " pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.511410 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gxdw\" (UniqueName: \"kubernetes.io/projected/e742a861-00e7-4ddc-b9d9-ab5dd22c84e2-kube-api-access-6gxdw\") pod \"service-ca-9c57cc56f-8v7wv\" (UID: \"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.529049 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:44 crc kubenswrapper[4936]: E0320 16:02:44.531921 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.031875066 +0000 UTC m=+115.978242881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.532159 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.540265 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dzgl2" event={"ID":"bf0f9e97-bddb-4054-99dd-d10428233e9f","Type":"ContainerStarted","Data":"e3512c43984200e291b8f5ecb4d15e19649ac93d93cfec43edf7f65e98290126"} Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.542435 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgcgg\" (UniqueName: \"kubernetes.io/projected/2c11e87a-9743-40b0-8741-dfec0ce96cb3-kube-api-access-jgcgg\") pod \"package-server-manager-789f6589d5-wspgg\" (UID: \"2c11e87a-9743-40b0-8741-dfec0ce96cb3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.543675 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" event={"ID":"7d71d251-484c-4198-b697-298f2d92031f","Type":"ContainerStarted","Data":"6f5c4ae3b227418af2f0a4a228301628047bb3a00101196166274f13f8ad6a23"} Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.543709 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" event={"ID":"7d71d251-484c-4198-b697-298f2d92031f","Type":"ContainerStarted","Data":"2ae6a811e038d95d5443c621e12133b49b1b4fe1e509846bd04eea924d67ef96"} Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.545158 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" event={"ID":"9333cf9d-b80d-4242-a940-c92ed3b593d1","Type":"ContainerStarted","Data":"edb98d9b986602424f3f6097e9cb7ebe0c6da2e355696f64ca8bc1b3100b78c8"} Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.548892 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" event={"ID":"cf423681-0ee6-4d44-bbed-e1447eaa7edd","Type":"ContainerStarted","Data":"a8d55d0f7fdf701493a1358e97332b9166940972302c1479cbe790c6c0a9aac3"} Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.550005 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.554576 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhp77\" (UniqueName: \"kubernetes.io/projected/e84e4377-e478-4c12-85bf-b236885152b2-kube-api-access-rhp77\") pod \"collect-profiles-29567040-9g8q5\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.568696 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" event={"ID":"ab39813c-c2d7-432c-aa05-7b33c1e86fac","Type":"ContainerStarted","Data":"df16c32748a31661bfdf4facd428473b6c6913c2602b8fa785b82ef4edc30bb3"} Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.579593 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62xb4\" (UniqueName: \"kubernetes.io/projected/6b4d3e08-82bf-4642-8145-cb6b1db82ad6-kube-api-access-62xb4\") pod \"machine-config-server-sc29d\" (UID: \"6b4d3e08-82bf-4642-8145-cb6b1db82ad6\") " pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.593416 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnvgp\" (UniqueName: \"kubernetes.io/projected/1766c31e-3eb1-4723-9fa9-4768f826f8ec-kube-api-access-vnvgp\") pod \"router-default-5444994796-9ltqh\" (UID: \"1766c31e-3eb1-4723-9fa9-4768f826f8ec\") " pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.605391 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74a1703c-e83e-487e-9d98-f2d948e3f831-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dbjm7\" (UID: \"74a1703c-e83e-487e-9d98-f2d948e3f831\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.631225 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: E0320 16:02:44.632006 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.131987474 +0000 UTC m=+116.078355289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.654123 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zprt6\" (UniqueName: \"kubernetes.io/projected/a0abe164-30be-4265-b4b2-4ce8c7b24597-kube-api-access-zprt6\") pod \"ingress-canary-qbm7q\" (UID: \"a0abe164-30be-4265-b4b2-4ce8c7b24597\") " pod="openshift-ingress-canary/ingress-canary-qbm7q" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.662373 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.663394 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-748h8\" (UniqueName: \"kubernetes.io/projected/aad2185e-c38e-4b5e-9253-5499237e913c-kube-api-access-748h8\") pod \"service-ca-operator-777779d784-nsh5l\" (UID: \"aad2185e-c38e-4b5e-9253-5499237e913c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.674256 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kxn4\" (UniqueName: \"kubernetes.io/projected/b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4-kube-api-access-6kxn4\") pod \"kube-storage-version-migrator-operator-b67b599dd-5l6qv\" (UID: \"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.720038 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95ed539e-3d39-44e4-a7c9-15f51fae30ae-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tvs2g\" (UID: \"95ed539e-3d39-44e4-a7c9-15f51fae30ae\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.732044 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.733650 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8v6h\" (UniqueName: \"kubernetes.io/projected/b0a7511a-0072-4933-bd7f-42f1203900ae-kube-api-access-j8v6h\") pod \"olm-operator-6b444d44fb-l96sz\" (UID: \"b0a7511a-0072-4933-bd7f-42f1203900ae\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: E0320 16:02:44.735046 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.235016072 +0000 UTC m=+116.181383887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.739193 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68cjb\" (UniqueName: \"kubernetes.io/projected/2d469c13-4b04-4c81-9b1b-177250681db4-kube-api-access-68cjb\") pod \"csi-hostpathplugin-dgtvg\" (UID: \"2d469c13-4b04-4c81-9b1b-177250681db4\") " pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.781361 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.788969 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.803170 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.838066 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.849028 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.850265 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.850987 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.851216 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.851306 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: E0320 16:02:44.851642 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.351629102 +0000 UTC m=+116.297996917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.855976 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qbm7q" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.864047 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sc29d" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.876870 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.885360 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dkllp"] Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.894014 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b65a2c6b-8b39-4742-bd96-6c4cbdc768ac-metrics-certs\") pod \"network-metrics-daemon-7jnpn\" (UID: \"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac\") " pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.906066 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.916989 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7jnpn" Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.954590 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:44 crc kubenswrapper[4936]: E0320 16:02:44.954812 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.454789284 +0000 UTC m=+116.401157089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:44 crc kubenswrapper[4936]: I0320 16:02:44.955198 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:44 crc kubenswrapper[4936]: E0320 16:02:44.955585 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.455577006 +0000 UTC m=+116.401944821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.057252 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.057950 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.557905155 +0000 UTC m=+116.504272970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: W0320 16:02:45.105014 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8498b530_52d4_4ca5_88ca_fda75dfbb18a.slice/crio-7f279364d5b8da80a29220ca7fda013d753e41ca195df8daa55aab47079dff55 WatchSource:0}: Error finding container 7f279364d5b8da80a29220ca7fda013d753e41ca195df8daa55aab47079dff55: Status 404 returned error can't find the container with id 7f279364d5b8da80a29220ca7fda013d753e41ca195df8daa55aab47079dff55 Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.164478 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.164982 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.664963334 +0000 UTC m=+116.611331149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.266945 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.267494 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.767472498 +0000 UTC m=+116.713840313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.381213 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.382046 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.882025342 +0000 UTC m=+116.828393157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.484413 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.484921 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.984864566 +0000 UTC m=+116.931232391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.485103 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.485698 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:45.985679668 +0000 UTC m=+116.932047483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.589825 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9ltqh" event={"ID":"1766c31e-3eb1-4723-9fa9-4768f826f8ec","Type":"ContainerStarted","Data":"67d7a399e487edf552dd56c9e17c8b8162f7959efd26396b8009bee8c46959d9"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.614592 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.615359 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.115332695 +0000 UTC m=+117.061700510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.660529 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" event={"ID":"b6ffb849-1700-439d-b7d3-1c1540a48fc1","Type":"ContainerStarted","Data":"ddd4b71f17de0a023a8956e5a02e42bd874827f4409038e0bb1cf61ebac3a898"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.661121 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" event={"ID":"b6ffb849-1700-439d-b7d3-1c1540a48fc1","Type":"ContainerStarted","Data":"b1a1058f47b79bb5fcaaa75ecd1bfeedd971cf1551b184c7ea855ed800cf5b65"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.663086 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.689920 4936 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2w76v container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.689998 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" podUID="b6ffb849-1700-439d-b7d3-1c1540a48fc1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.698820 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" event={"ID":"1f1994bb-a56f-4e25-b6d8-7bcc2113de10","Type":"ContainerStarted","Data":"73d783166ed1ed96266035aea8bfbda72f9dc7a7e399ab1ec69b67b1998b24ad"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.704983 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-ssr5n" podStartSLOduration=54.704941286 podStartE2EDuration="54.704941286s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:45.704666249 +0000 UTC m=+116.651034064" watchObservedRunningTime="2026-03-20 16:02:45.704941286 +0000 UTC m=+116.651309121" Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.706418 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" event={"ID":"9333cf9d-b80d-4242-a940-c92ed3b593d1","Type":"ContainerStarted","Data":"c7e27bbd84991bdee07cd35e5752be306d3cd838ff6bdddba0e68de90443db8b"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.707002 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.709015 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sc29d" event={"ID":"6b4d3e08-82bf-4642-8145-cb6b1db82ad6","Type":"ContainerStarted","Data":"92b0bd92eeecf19d8a79a8bd860fab4417b1c785d888d21ce734f3ce10f19dd5"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.716732 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.717123 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.217108399 +0000 UTC m=+117.163476214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.726266 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" event={"ID":"ab39813c-c2d7-432c-aa05-7b33c1e86fac","Type":"ContainerStarted","Data":"57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.727035 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.731681 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dzgl2" event={"ID":"bf0f9e97-bddb-4054-99dd-d10428233e9f","Type":"ContainerStarted","Data":"19f0314b694b7fff0279da4136e2dc94fab0e00ee2152878225659d5199afedb"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.732499 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-dzgl2" Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.743678 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" event={"ID":"6376c9a4-9209-4f0d-835d-7db09de15649","Type":"ContainerStarted","Data":"44f3622d8fd3bb32d93283d54a4590e5cd33789d585f501edb2a965eefd785d4"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.743996 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" event={"ID":"6376c9a4-9209-4f0d-835d-7db09de15649","Type":"ContainerStarted","Data":"896bd4fb44c21e49c902e95ca9ba7aca7286028adf13b73abfb0c475a73ea38f"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.748801 4936 patch_prober.go:28] interesting pod/downloads-7954f5f757-dzgl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.748978 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dzgl2" podUID="bf0f9e97-bddb-4054-99dd-d10428233e9f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.748810 4936 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jhqvm container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" start-of-body= Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.749270 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" podUID="9333cf9d-b80d-4242-a940-c92ed3b593d1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.775028 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" event={"ID":"8498b530-52d4-4ca5-88ca-fda75dfbb18a","Type":"ContainerStarted","Data":"7f279364d5b8da80a29220ca7fda013d753e41ca195df8daa55aab47079dff55"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.778041 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" event={"ID":"7d71d251-484c-4198-b697-298f2d92031f","Type":"ContainerStarted","Data":"528f87ffe82ce97c78a4d06c85230816d88317e0f2bab96d58aa947792a63325"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.782387 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" event={"ID":"cf423681-0ee6-4d44-bbed-e1447eaa7edd","Type":"ContainerStarted","Data":"ca77cf0de3bc191c26650465dd655ee306874be51f508505b9880735dc0b2f68"} Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.818170 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.821494 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.321467414 +0000 UTC m=+117.267835229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.920711 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:45 crc kubenswrapper[4936]: E0320 16:02:45.921266 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.421248384 +0000 UTC m=+117.367616199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:45 crc kubenswrapper[4936]: I0320 16:02:45.986868 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" podStartSLOduration=54.986841388 podStartE2EDuration="54.986841388s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:45.983761524 +0000 UTC m=+116.930129349" watchObservedRunningTime="2026-03-20 16:02:45.986841388 +0000 UTC m=+116.933209203" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.023329 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.023863 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.52382389 +0000 UTC m=+117.470191705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.032226 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.035051 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.124990 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.125844 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.62581117 +0000 UTC m=+117.572178985 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.229783 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.230229 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.730207946 +0000 UTC m=+117.676575761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.253729 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" podStartSLOduration=55.253702539 podStartE2EDuration="55.253702539s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.251822778 +0000 UTC m=+117.198190593" watchObservedRunningTime="2026-03-20 16:02:46.253702539 +0000 UTC m=+117.200070354" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.274924 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.292364 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vqmvx" podStartSLOduration=55.292343026 podStartE2EDuration="55.292343026s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.290011592 +0000 UTC m=+117.236379427" watchObservedRunningTime="2026-03-20 16:02:46.292343026 +0000 UTC m=+117.238710841" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.331908 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k98jk" podStartSLOduration=55.331881658 podStartE2EDuration="55.331881658s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.320729073 +0000 UTC m=+117.267096908" watchObservedRunningTime="2026-03-20 16:02:46.331881658 +0000 UTC m=+117.278249473" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.333740 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.336857 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.836839543 +0000 UTC m=+117.783207358 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.365754 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-dzgl2" podStartSLOduration=55.365730434 podStartE2EDuration="55.365730434s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.360168222 +0000 UTC m=+117.306536037" watchObservedRunningTime="2026-03-20 16:02:46.365730434 +0000 UTC m=+117.312098249" Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.436069 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.936049647 +0000 UTC m=+117.882417462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.435979 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.436396 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.436843 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:46.936833749 +0000 UTC m=+117.883201564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.455933 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" podStartSLOduration=55.455908261 podStartE2EDuration="55.455908261s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.406854769 +0000 UTC m=+117.353222594" watchObservedRunningTime="2026-03-20 16:02:46.455908261 +0000 UTC m=+117.402276076" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.489466 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" podStartSLOduration=5.489435298 podStartE2EDuration="5.489435298s" podCreationTimestamp="2026-03-20 16:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.488335418 +0000 UTC m=+117.434703233" watchObservedRunningTime="2026-03-20 16:02:46.489435298 +0000 UTC m=+117.435803113" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.535057 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.537624 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.538007 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.037987856 +0000 UTC m=+117.984355671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: W0320 16:02:46.572798 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90dd52f0_03cd_46fd_90cd_1a480bae1280.slice/crio-6e207e6d63448f59fda3cecc94461cbf4bc2ad609ecbad30a9a8a2e77a5916d7 WatchSource:0}: Error finding container 6e207e6d63448f59fda3cecc94461cbf4bc2ad609ecbad30a9a8a2e77a5916d7: Status 404 returned error can't find the container with id 6e207e6d63448f59fda3cecc94461cbf4bc2ad609ecbad30a9a8a2e77a5916d7 Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.577431 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r4w72"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.597770 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.599568 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.603647 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6q5fl"] Mar 20 16:02:46 crc kubenswrapper[4936]: W0320 16:02:46.606086 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod151dc193_527a_4506_bcb9_170641b9ea3f.slice/crio-2089451c9589a05141b8cb6cb9cf38827c3c5867b82ee8a31a8a21def880da33 WatchSource:0}: Error finding container 2089451c9589a05141b8cb6cb9cf38827c3c5867b82ee8a31a8a21def880da33: Status 404 returned error can't find the container with id 2089451c9589a05141b8cb6cb9cf38827c3c5867b82ee8a31a8a21def880da33 Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.620157 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jgfjb"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.623065 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl"] Mar 20 16:02:46 crc kubenswrapper[4936]: W0320 16:02:46.627123 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e7175b7_4b92_49c8_98d1_b2366c099ced.slice/crio-2281f3b423cc3082727950173906a62283c9bdd30fc75b6c38cf9dcfcd47b160 WatchSource:0}: Error finding container 2281f3b423cc3082727950173906a62283c9bdd30fc75b6c38cf9dcfcd47b160: Status 404 returned error can't find the container with id 2281f3b423cc3082727950173906a62283c9bdd30fc75b6c38cf9dcfcd47b160 Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.639005 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.639373 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.139358349 +0000 UTC m=+118.085726164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.709129 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.743183 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.743674 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.243642522 +0000 UTC m=+118.190010347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: W0320 16:02:46.766442 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cd73a43_1689_479b_bae3_6519bdd522b5.slice/crio-2baaebecbae5e2ebf6804a54c0ba9f6a22480b064ad83bec269a047efbf43dcc WatchSource:0}: Error finding container 2baaebecbae5e2ebf6804a54c0ba9f6a22480b064ad83bec269a047efbf43dcc: Status 404 returned error can't find the container with id 2baaebecbae5e2ebf6804a54c0ba9f6a22480b064ad83bec269a047efbf43dcc Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.793211 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.801256 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6" event={"ID":"1cd73a43-1689-479b-bae3-6519bdd522b5","Type":"ContainerStarted","Data":"2baaebecbae5e2ebf6804a54c0ba9f6a22480b064ad83bec269a047efbf43dcc"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.824227 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5"] Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.839237 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" event={"ID":"151dc193-527a-4506-bcb9-170641b9ea3f","Type":"ContainerStarted","Data":"2089451c9589a05141b8cb6cb9cf38827c3c5867b82ee8a31a8a21def880da33"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.845702 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.846697 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.346681061 +0000 UTC m=+118.293048876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.848150 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" event={"ID":"242b082c-0e13-4b74-830e-0bb1008b5c0a","Type":"ContainerStarted","Data":"8dd0688728d7fba64e6f28578e339e5258b5449c80bbbc63baf6d275da9251d2"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.848185 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" event={"ID":"242b082c-0e13-4b74-830e-0bb1008b5c0a","Type":"ContainerStarted","Data":"76a9757a38713ef5bc8078f39741d7495780b1cc97e9e0b224a0a9d80f41d67c"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.849464 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.851780 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6q5fl" event={"ID":"c52c0329-ca68-4d95-8341-ab8191c3a186","Type":"ContainerStarted","Data":"225ff718e9e7e6534de7e30034b9c3c82fa1cb352430832486bc8078bd1791b4"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.853604 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" event={"ID":"967b1700-c689-4195-8dcd-4177b702d066","Type":"ContainerStarted","Data":"9ec274b6dffceda00c68ab6638c39b25d361d3bc62936bfae2c51efaf441ed95"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.855872 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" event={"ID":"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504","Type":"ContainerStarted","Data":"1815c422119396d126fb12957d5853367d078ab3203e63d00aaa3f6fe6e72b58"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.855896 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" event={"ID":"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504","Type":"ContainerStarted","Data":"c17b207b1bee69b8e1d41f94e4594e973a8e6023f2948b438b7c81abf655c81e"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.855906 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" event={"ID":"50acf6a8-8d00-4bd2-9a88-f85cd8b4b504","Type":"ContainerStarted","Data":"6ecd2b3238b41bbf60a520224e3fe41b4a7c6405b6c17af613fd0d16acd1d81c"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.872602 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9ltqh" event={"ID":"1766c31e-3eb1-4723-9fa9-4768f826f8ec","Type":"ContainerStarted","Data":"74a25e639fa9f5cf96133ad6a515e408b1f593ffbf9b53ab65eff892fe63b8ae"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.893686 4936 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-ccjqw container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.893758 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" podUID="242b082c-0e13-4b74-830e-0bb1008b5c0a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.913593 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" podStartSLOduration=55.913566621 podStartE2EDuration="55.913566621s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.91023267 +0000 UTC m=+117.856600485" watchObservedRunningTime="2026-03-20 16:02:46.913566621 +0000 UTC m=+117.859934426" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.923729 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" event={"ID":"90dd52f0-03cd-46fd-90cd-1a480bae1280","Type":"ContainerStarted","Data":"6e207e6d63448f59fda3cecc94461cbf4bc2ad609ecbad30a9a8a2e77a5916d7"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.933710 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sc29d" event={"ID":"6b4d3e08-82bf-4642-8145-cb6b1db82ad6","Type":"ContainerStarted","Data":"bc7261f37abd852b9b66041f1db5ca2bb03f322ec52bba4a97baf61fa0482b94"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.944331 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-9ltqh" podStartSLOduration=55.944310982 podStartE2EDuration="55.944310982s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.937056363 +0000 UTC m=+117.883424178" watchObservedRunningTime="2026-03-20 16:02:46.944310982 +0000 UTC m=+117.890678797" Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.948748 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.949057 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.44899836 +0000 UTC m=+118.395366185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.949197 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:46 crc kubenswrapper[4936]: E0320 16:02:46.955728 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.455684783 +0000 UTC m=+118.402052598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.974826 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" event={"ID":"7e7175b7-4b92-49c8-98d1-b2366c099ced","Type":"ContainerStarted","Data":"2281f3b423cc3082727950173906a62283c9bdd30fc75b6c38cf9dcfcd47b160"} Mar 20 16:02:46 crc kubenswrapper[4936]: I0320 16:02:46.984622 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nnnkc" podStartSLOduration=55.984591454 podStartE2EDuration="55.984591454s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.963154527 +0000 UTC m=+117.909522332" watchObservedRunningTime="2026-03-20 16:02:46.984591454 +0000 UTC m=+117.930959269" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:46.996188 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" event={"ID":"0f7a4f62-f974-487f-90d0-f45d16638c4c","Type":"ContainerStarted","Data":"b4390b89058f3c29ab83f10645ff64220401ea468c728192336a0baf81debdba"} Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:46.996258 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" event={"ID":"0f7a4f62-f974-487f-90d0-f45d16638c4c","Type":"ContainerStarted","Data":"e4737cdc9b077beaae02de82c6dfb17543ce75c0e100d5c3fd7e56fad7773a27"} Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:46.999754 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-sc29d" podStartSLOduration=5.999735358 podStartE2EDuration="5.999735358s" podCreationTimestamp="2026-03-20 16:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:46.995815881 +0000 UTC m=+117.942183696" watchObservedRunningTime="2026-03-20 16:02:46.999735358 +0000 UTC m=+117.946103173" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.001706 4936 generic.go:334] "Generic (PLEG): container finished" podID="6376c9a4-9209-4f0d-835d-7db09de15649" containerID="44f3622d8fd3bb32d93283d54a4590e5cd33789d585f501edb2a965eefd785d4" exitCode=0 Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.001789 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" event={"ID":"6376c9a4-9209-4f0d-835d-7db09de15649","Type":"ContainerDied","Data":"44f3622d8fd3bb32d93283d54a4590e5cd33789d585f501edb2a965eefd785d4"} Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.009224 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" event={"ID":"1f1994bb-a56f-4e25-b6d8-7bcc2113de10","Type":"ContainerStarted","Data":"2c46efedf8c9114e10f698eb811dff9cbb3d96b6c78ec1eaea16d81b38ce96fb"} Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.009261 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" event={"ID":"1f1994bb-a56f-4e25-b6d8-7bcc2113de10","Type":"ContainerStarted","Data":"6d2cc18830bfc2e51a58b5152b280ecfa7557ab7bafbe5b41104123a0eb95ddb"} Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.015446 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" event={"ID":"d3739ff0-d3fa-4243-9529-209e3ec9e4e4","Type":"ContainerStarted","Data":"4f5a5b84bc720f1b3933512c86d90d7845c5ba5340b66137bd4cf5e4ca7d4cec"} Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.019250 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" event={"ID":"32e4b213-a66b-4e97-83b1-cac96a741b9a","Type":"ContainerStarted","Data":"d0e42dcc5d572f5ca99d04d3f399587fb8586f699fdfdfb90edeaedb96560604"} Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.025355 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" event={"ID":"8498b530-52d4-4ca5-88ca-fda75dfbb18a","Type":"ContainerStarted","Data":"7620064e0f4b3d4db6fddb6416578872616d7de6d5c1c27dc1aaca7a9de19dac"} Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.040431 4936 patch_prober.go:28] interesting pod/downloads-7954f5f757-dzgl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.040507 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dzgl2" podUID="bf0f9e97-bddb-4054-99dd-d10428233e9f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.053436 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.056174 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.556143691 +0000 UTC m=+118.502511506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.058647 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.065069 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.069664 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.073701 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.573680351 +0000 UTC m=+118.520048166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.096926 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.107561 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-q6cx6" podStartSLOduration=56.107515267 podStartE2EDuration="56.107515267s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:47.068022656 +0000 UTC m=+118.014390471" watchObservedRunningTime="2026-03-20 16:02:47.107515267 +0000 UTC m=+118.053883092" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.139534 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xtqwz"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.147083 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7jnpn"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.159238 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.165753 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.168496 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.169439 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qbm7q"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.170509 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.170797 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.670765747 +0000 UTC m=+118.617133562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.172089 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.174731 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.180746 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.187067 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.189682 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j54jw"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.193646 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xlnmq"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.198991 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8v7wv"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.204982 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dgtvg"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.206509 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dkllp" podStartSLOduration=56.206491084 podStartE2EDuration="56.206491084s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:47.139341337 +0000 UTC m=+118.085709152" watchObservedRunningTime="2026-03-20 16:02:47.206491084 +0000 UTC m=+118.152858899" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.219019 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5nbl4"] Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.227732 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg"] Mar 20 16:02:47 crc kubenswrapper[4936]: W0320 16:02:47.245107 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a005177_344e_461b_b8ac_989f18ade02d.slice/crio-c5e6adeae968c9f929c988b4f5785d43e9daa2e4901b391e6a442e1595eba445 WatchSource:0}: Error finding container c5e6adeae968c9f929c988b4f5785d43e9daa2e4901b391e6a442e1595eba445: Status 404 returned error can't find the container with id c5e6adeae968c9f929c988b4f5785d43e9daa2e4901b391e6a442e1595eba445 Mar 20 16:02:47 crc kubenswrapper[4936]: W0320 16:02:47.254389 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74a1703c_e83e_487e_9d98_f2d948e3f831.slice/crio-01a421809ba45537d130fc88e630cf717b2a216eaba411fefc0a638510e61338 WatchSource:0}: Error finding container 01a421809ba45537d130fc88e630cf717b2a216eaba411fefc0a638510e61338: Status 404 returned error can't find the container with id 01a421809ba45537d130fc88e630cf717b2a216eaba411fefc0a638510e61338 Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.276789 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.277164 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.777149097 +0000 UTC m=+118.723516912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: W0320 16:02:47.287631 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb1e412b_f423_4178_a306_b9201f089099.slice/crio-69afcf94f61f2ae6200f523c940cfdb495355a2bd418a283e4b356c9d2c0692b WatchSource:0}: Error finding container 69afcf94f61f2ae6200f523c940cfdb495355a2bd418a283e4b356c9d2c0692b: Status 404 returned error can't find the container with id 69afcf94f61f2ae6200f523c940cfdb495355a2bd418a283e4b356c9d2c0692b Mar 20 16:02:47 crc kubenswrapper[4936]: W0320 16:02:47.339898 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb08d05f0_ce9d_49b5_b6ed_f2d02e73b1f4.slice/crio-4ea5a4faa5e46473dd8262caab220afa9191953c197df4d5986308a58815c415 WatchSource:0}: Error finding container 4ea5a4faa5e46473dd8262caab220afa9191953c197df4d5986308a58815c415: Status 404 returned error can't find the container with id 4ea5a4faa5e46473dd8262caab220afa9191953c197df4d5986308a58815c415 Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.377117 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.377349 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.877318858 +0000 UTC m=+118.823686673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.377980 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.378359 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.878345316 +0000 UTC m=+118.824713121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.428053 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2w76v" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.495977 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.496513 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:47.996496308 +0000 UTC m=+118.942864123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.597666 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.597947 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.097929923 +0000 UTC m=+119.044297728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.729800 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.730851 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.230828019 +0000 UTC m=+119.177195824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.835249 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.835668 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.335652806 +0000 UTC m=+119.282020621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.850671 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.863787 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:47 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:47 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:47 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.863885 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.936639 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:47 crc kubenswrapper[4936]: E0320 16:02:47.937109 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.437085441 +0000 UTC m=+119.383453256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:47 crc kubenswrapper[4936]: I0320 16:02:47.973348 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-m5rqk"] Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.038555 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.039214 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.539199715 +0000 UTC m=+119.485567520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.045447 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" event={"ID":"cd07f1f3-ea79-4612-af59-91151c9bb06f","Type":"ContainerStarted","Data":"244c79285794dcc114128dd18f9b8e672208052e65ea3dd21423bb834837f6d4"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.131656 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6q5fl" event={"ID":"c52c0329-ca68-4d95-8341-ab8191c3a186","Type":"ContainerStarted","Data":"6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.144900 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.145150 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.645116852 +0000 UTC m=+119.591484677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.145311 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.145701 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.645691868 +0000 UTC m=+119.592059883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.156045 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" event={"ID":"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4","Type":"ContainerStarted","Data":"4ea5a4faa5e46473dd8262caab220afa9191953c197df4d5986308a58815c415"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.169817 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" event={"ID":"fb1e412b-f423-4178-a306-b9201f089099","Type":"ContainerStarted","Data":"69afcf94f61f2ae6200f523c940cfdb495355a2bd418a283e4b356c9d2c0692b"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.196085 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" event={"ID":"676cb1dc-3786-44f8-bf27-963da433b3c1","Type":"ContainerStarted","Data":"a7386d6e50a64df4a14f85d5c9a73748452965b388753cff5f76c847ada3209c"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.197971 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7jnpn" event={"ID":"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac","Type":"ContainerStarted","Data":"bb06ccac9f94b3548de213860bc5c506d78f45b4a2dd6e7b58dba8607e8c10c3"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.200318 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" event={"ID":"95ed539e-3d39-44e4-a7c9-15f51fae30ae","Type":"ContainerStarted","Data":"3df5a57bef3bf9e1d4c1e5a9d8897cfe7b4adec3b694d4c5bb31ae9ddbdb6196"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.204353 4936 generic.go:334] "Generic (PLEG): container finished" podID="151dc193-527a-4506-bcb9-170641b9ea3f" containerID="266191bf444f8ca0c46bfb9cbf48721a39f8191918a311760978ea27f5410241" exitCode=0 Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.205696 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" event={"ID":"151dc193-527a-4506-bcb9-170641b9ea3f","Type":"ContainerDied","Data":"266191bf444f8ca0c46bfb9cbf48721a39f8191918a311760978ea27f5410241"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.237810 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-6q5fl" podStartSLOduration=57.237780737 podStartE2EDuration="57.237780737s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.173279693 +0000 UTC m=+119.119647508" watchObservedRunningTime="2026-03-20 16:02:48.237780737 +0000 UTC m=+119.184148552" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.246298 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.247393 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.74737094 +0000 UTC m=+119.693738755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.257431 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" event={"ID":"e84e4377-e478-4c12-85bf-b236885152b2","Type":"ContainerStarted","Data":"110680cf2e9247bb1db2afd5997a349c5a6285e2bf3818bb3fa35e00d004f836"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.257926 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" event={"ID":"e84e4377-e478-4c12-85bf-b236885152b2","Type":"ContainerStarted","Data":"e361365e6b1edbc6ddd5ab373f6dc57b182135e658f87174c7167d0ee85d4090"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.287052 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" event={"ID":"32e4b213-a66b-4e97-83b1-cac96a741b9a","Type":"ContainerStarted","Data":"fda7e6c2bed66f8f173cc1b4cfa8f4e6a0815c917b06f4fc27f6ce427ac81f2d"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.319608 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" podStartSLOduration=57.319581524 podStartE2EDuration="57.319581524s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.318617368 +0000 UTC m=+119.264985183" watchObservedRunningTime="2026-03-20 16:02:48.319581524 +0000 UTC m=+119.265949339" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.333754 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" event={"ID":"9027bafb-1a39-4a53-9bcf-778b8859c192","Type":"ContainerStarted","Data":"e320e54aaceff90ff60e68ce3f9d166b51dc596c0008600b92026fd635d4762d"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.361461 4936 generic.go:334] "Generic (PLEG): container finished" podID="7e7175b7-4b92-49c8-98d1-b2366c099ced" containerID="b4cdc2455a4e955a13ba9e3cf78626f4d45acc941ca68c3e5ab930e0dbe12fb3" exitCode=0 Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.361567 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" event={"ID":"7e7175b7-4b92-49c8-98d1-b2366c099ced","Type":"ContainerDied","Data":"b4cdc2455a4e955a13ba9e3cf78626f4d45acc941ca68c3e5ab930e0dbe12fb3"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.362690 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.363990 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.863972019 +0000 UTC m=+119.810339834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.368462 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6" event={"ID":"1cd73a43-1689-479b-bae3-6519bdd522b5","Type":"ContainerStarted","Data":"138a140879a60c7192b9f2e1f5c61080d540033255a870d02e9894e23fd6c352"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.376903 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xtqwz" event={"ID":"3a005177-344e-461b-b8ac-989f18ade02d","Type":"ContainerStarted","Data":"c5e6adeae968c9f929c988b4f5785d43e9daa2e4901b391e6a442e1595eba445"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.390170 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-64r5t" podStartSLOduration=57.390146115 podStartE2EDuration="57.390146115s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.364982106 +0000 UTC m=+119.311349921" watchObservedRunningTime="2026-03-20 16:02:48.390146115 +0000 UTC m=+119.336513930" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.458936 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" event={"ID":"967b1700-c689-4195-8dcd-4177b702d066","Type":"ContainerStarted","Data":"7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.460040 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.464852 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.466381 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:48.96635838 +0000 UTC m=+119.912726195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.482960 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" event={"ID":"74a1703c-e83e-487e-9d98-f2d948e3f831","Type":"ContainerStarted","Data":"01a421809ba45537d130fc88e630cf717b2a216eaba411fefc0a638510e61338"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.498992 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" podStartSLOduration=57.498969292 podStartE2EDuration="57.498969292s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.498391676 +0000 UTC m=+119.444759501" watchObservedRunningTime="2026-03-20 16:02:48.498969292 +0000 UTC m=+119.445337107" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.504476 4936 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r4w72 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.504592 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" podUID="967b1700-c689-4195-8dcd-4177b702d066" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.521296 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" event={"ID":"5d3858e7-c19f-4140-847c-7742c0af0d9c","Type":"ContainerStarted","Data":"67942bacd6f0ec121f3cf5e4fa0c17324c83362bc1d1841527dcf9b45aad4b1a"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.521365 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" event={"ID":"5d3858e7-c19f-4140-847c-7742c0af0d9c","Type":"ContainerStarted","Data":"3674b386f325d248d93177bb50837bfa11c54eeebc1c4dfd563b53514b5d695f"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.527520 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" event={"ID":"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2","Type":"ContainerStarted","Data":"6a14dc2487e406f0d2f7031c2f88e3374724ac2819ca449496142ea801f26f5c"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.527585 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" event={"ID":"e742a861-00e7-4ddc-b9d9-ab5dd22c84e2","Type":"ContainerStarted","Data":"38195bd99f58bfece3e197f59d1d0c5fe881ee717cc8225b46dbad98b6d16256"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.563362 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-slzkt" podStartSLOduration=57.563332853 podStartE2EDuration="57.563332853s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.562274754 +0000 UTC m=+119.508642569" watchObservedRunningTime="2026-03-20 16:02:48.563332853 +0000 UTC m=+119.509700668" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.566818 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.583954 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" event={"ID":"b0a7511a-0072-4933-bd7f-42f1203900ae","Type":"ContainerStarted","Data":"29a1f599341cfb54695e7119c34d28a9f23aeb02261ae9db1a58c3bd3546c216"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.584031 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" event={"ID":"b0a7511a-0072-4933-bd7f-42f1203900ae","Type":"ContainerStarted","Data":"4fa064b41d1d78a88d5729861dfe0ed164771b85c64d05e774e35b6e1d327bde"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.592039 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.592153 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.09212463 +0000 UTC m=+120.038492445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.593357 4936 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-l96sz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.593424 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" podUID="b0a7511a-0072-4933-bd7f-42f1203900ae" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.633124 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" event={"ID":"90dd52f0-03cd-46fd-90cd-1a480bae1280","Type":"ContainerStarted","Data":"22ae6722f7270c08b4a5e885726a3dad97163b01509087f30d4801a21665a757"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.638994 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-8v7wv" podStartSLOduration=57.638966992 podStartE2EDuration="57.638966992s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.636886845 +0000 UTC m=+119.583254660" watchObservedRunningTime="2026-03-20 16:02:48.638966992 +0000 UTC m=+119.585334807" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.658734 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56034: no serving certificate available for the kubelet" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.685379 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.685836 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.185796733 +0000 UTC m=+120.132164548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.687359 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" event={"ID":"d3739ff0-d3fa-4243-9529-209e3ec9e4e4","Type":"ContainerStarted","Data":"58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.687879 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.701536 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xxvbh" podStartSLOduration=57.701511353 podStartE2EDuration="57.701511353s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.698270194 +0000 UTC m=+119.644638009" watchObservedRunningTime="2026-03-20 16:02:48.701511353 +0000 UTC m=+119.647879168" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.734703 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qbm7q" event={"ID":"a0abe164-30be-4265-b4b2-4ce8c7b24597","Type":"ContainerStarted","Data":"0e368e224e441060d82ccb28d3cdf58f4b585ff5cd2aaadefab06388f603487a"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.763091 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56050: no serving certificate available for the kubelet" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.763945 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" podStartSLOduration=57.76392082 podStartE2EDuration="57.76392082s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.763277413 +0000 UTC m=+119.709645228" watchObservedRunningTime="2026-03-20 16:02:48.76392082 +0000 UTC m=+119.710288625" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.764164 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" podStartSLOduration=57.764159557 podStartE2EDuration="57.764159557s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.728118751 +0000 UTC m=+119.674486566" watchObservedRunningTime="2026-03-20 16:02:48.764159557 +0000 UTC m=+119.710527372" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.788723 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.790704 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.290684102 +0000 UTC m=+120.237051917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.800418 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" event={"ID":"2c11e87a-9743-40b0-8741-dfec0ce96cb3","Type":"ContainerStarted","Data":"7fd22e218810d358597030a04bca5ec42c3c5038d339b2e635213b42091966ae"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.803136 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qbm7q" podStartSLOduration=7.803114302 podStartE2EDuration="7.803114302s" podCreationTimestamp="2026-03-20 16:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.803114592 +0000 UTC m=+119.749482427" watchObservedRunningTime="2026-03-20 16:02:48.803114302 +0000 UTC m=+119.749482117" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.812161 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" event={"ID":"aad2185e-c38e-4b5e-9253-5499237e913c","Type":"ContainerStarted","Data":"2c10cc2527431815d2055fc1204e4b52ca096a37cdca34c2c80680c1f0824e6b"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.812230 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" event={"ID":"aad2185e-c38e-4b5e-9253-5499237e913c","Type":"ContainerStarted","Data":"81e99ec69bb904e60ea903c9d248c615c9c06ae6520cadf82648a441ba24a28d"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.848392 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" event={"ID":"6376c9a4-9209-4f0d-835d-7db09de15649","Type":"ContainerStarted","Data":"85bf34fc027a1bacfffed5a7fb98c9fd007b97b41b18fe02e7f3373fa7f80272"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.849257 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.856852 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nsh5l" podStartSLOduration=57.856822982 podStartE2EDuration="57.856822982s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.849522922 +0000 UTC m=+119.795890727" watchObservedRunningTime="2026-03-20 16:02:48.856822982 +0000 UTC m=+119.803190797" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.863288 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56060: no serving certificate available for the kubelet" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.878627 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" event={"ID":"0f7a4f62-f974-487f-90d0-f45d16638c4c","Type":"ContainerStarted","Data":"3d7d85b6027e57570b06b498bb674995f05fd0e780f866a57b70f41771dd18f7"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.896944 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:48 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:48 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:48 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.897031 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.898115 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:48 crc kubenswrapper[4936]: E0320 16:02:48.899284 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.399239902 +0000 UTC m=+120.345607717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.903807 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.937278 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p8jqw" podStartSLOduration=57.937261102 podStartE2EDuration="57.937261102s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.936889392 +0000 UTC m=+119.883257197" watchObservedRunningTime="2026-03-20 16:02:48.937261102 +0000 UTC m=+119.883628917" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.939237 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" podStartSLOduration=57.939232846 podStartE2EDuration="57.939232846s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:48.900563298 +0000 UTC m=+119.846931123" watchObservedRunningTime="2026-03-20 16:02:48.939232846 +0000 UTC m=+119.885600661" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.949133 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5nbl4" event={"ID":"921365e5-3846-485e-9afa-270834e81db5","Type":"ContainerStarted","Data":"7f6e76cb1efba88199bd12068eaaf7953f05745df6e11fc503087521d3d9ec8b"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.959079 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" event={"ID":"2d469c13-4b04-4c81-9b1b-177250681db4","Type":"ContainerStarted","Data":"8c1a6dfbf62e2b46c53873ed47ddb7bec3b7dfbc85baf189959d661faf3554a1"} Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.959290 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56064: no serving certificate available for the kubelet" Mar 20 16:02:48 crc kubenswrapper[4936]: I0320 16:02:48.983923 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" event={"ID":"4f8268de-4dd1-4521-9e5a-b4679996b511","Type":"ContainerStarted","Data":"672cc30bf2256470dc5280b2481311a3b5e6e087a1e815456c5484492e9ac717"} Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.002238 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ccjqw" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.002719 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.003248 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.503225967 +0000 UTC m=+120.449593772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.069341 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56078: no serving certificate available for the kubelet" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.104715 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.106450 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.60641142 +0000 UTC m=+120.552779235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.177936 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56088: no serving certificate available for the kubelet" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.207620 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.208016 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.708002509 +0000 UTC m=+120.654370324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.267194 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56092: no serving certificate available for the kubelet" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.311254 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.311517 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.811454149 +0000 UTC m=+120.757821964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.311789 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.312181 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.812162329 +0000 UTC m=+120.758530144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.385803 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56098: no serving certificate available for the kubelet" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.413915 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.414134 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.914101557 +0000 UTC m=+120.860469372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.414358 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.414779 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:49.914772766 +0000 UTC m=+120.861140581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.515962 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.516348 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.016330264 +0000 UTC m=+120.962698079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.623602 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.624060 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.124037941 +0000 UTC m=+121.070405746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.702131 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5cpcn"] Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.704469 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.717596 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.726852 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.727099 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.227043778 +0000 UTC m=+121.173411613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.727312 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.727801 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.227784989 +0000 UTC m=+121.174152804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.728387 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5cpcn"] Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.828069 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.828667 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-catalog-content\") pod \"certified-operators-5cpcn\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.828698 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-utilities\") pod \"certified-operators-5cpcn\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.828779 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd9t9\" (UniqueName: \"kubernetes.io/projected/7964eb91-819c-4853-a598-9b93346f496f-kube-api-access-vd9t9\") pod \"certified-operators-5cpcn\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.828895 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.328876794 +0000 UTC m=+121.275244609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.878923 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:49 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:49 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:49 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.878995 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.909289 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-km59s"] Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.910287 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.917806 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.938657 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-catalog-content\") pod \"certified-operators-5cpcn\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.938726 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-utilities\") pod \"certified-operators-5cpcn\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.938817 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.938856 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd9t9\" (UniqueName: \"kubernetes.io/projected/7964eb91-819c-4853-a598-9b93346f496f-kube-api-access-vd9t9\") pod \"certified-operators-5cpcn\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.939676 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-catalog-content\") pod \"certified-operators-5cpcn\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.939913 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-utilities\") pod \"certified-operators-5cpcn\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:49 crc kubenswrapper[4936]: E0320 16:02:49.940216 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.44019897 +0000 UTC m=+121.386566785 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.955800 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-km59s"] Mar 20 16:02:49 crc kubenswrapper[4936]: I0320 16:02:49.982479 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd9t9\" (UniqueName: \"kubernetes.io/projected/7964eb91-819c-4853-a598-9b93346f496f-kube-api-access-vd9t9\") pod \"certified-operators-5cpcn\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.023275 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" event={"ID":"74a1703c-e83e-487e-9d98-f2d948e3f831","Type":"ContainerStarted","Data":"4705306ed7c977378743eee88e73552a196daa802c9d0494c3d6f6e51f5238ca"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.023721 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" event={"ID":"74a1703c-e83e-487e-9d98-f2d948e3f831","Type":"ContainerStarted","Data":"fff9df1a7357fb1dfdec76715910ef0dad8191f3cff37c7160714492497cc3a0"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.032816 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" event={"ID":"cd07f1f3-ea79-4612-af59-91151c9bb06f","Type":"ContainerStarted","Data":"719f0790fbc2b447a80eecd74e84dc40072f9c627bd37f8e890afc7b2db8956d"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.040369 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.041050 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s5fs\" (UniqueName: \"kubernetes.io/projected/2aae37a3-f585-4b9b-94e3-5560e06da84c-kube-api-access-5s5fs\") pod \"community-operators-km59s\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.041077 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.541043869 +0000 UTC m=+121.487411684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.041379 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.041447 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-utilities\") pod \"community-operators-km59s\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.041479 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-catalog-content\") pod \"community-operators-km59s\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.041985 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.541964444 +0000 UTC m=+121.488332259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.044946 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" event={"ID":"fb1e412b-f423-4178-a306-b9201f089099","Type":"ContainerStarted","Data":"dd796d66cd407c6e35e2884e52210d22166d357a34b5dd296891537223865c54"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.045015 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" event={"ID":"fb1e412b-f423-4178-a306-b9201f089099","Type":"ContainerStarted","Data":"39b63d8ac67b4dc274449c5370b866d0da568583f568b67abeaff80c32233106"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.055085 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.057955 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.071569 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" event={"ID":"7e7175b7-4b92-49c8-98d1-b2366c099ced","Type":"ContainerStarted","Data":"1fa6084214754ce1b5a20ba00a8ed4ef6d8874861c6ae40489c2238f309baafa"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.093956 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" event={"ID":"95ed539e-3d39-44e4-a7c9-15f51fae30ae","Type":"ContainerStarted","Data":"3ef903d91cebb0b1a8602b01bb82704856812a5644d6e0725b6e3d8a5c6505bb"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.106267 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9fw5r"] Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.107401 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.129231 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" event={"ID":"151dc193-527a-4506-bcb9-170641b9ea3f","Type":"ContainerStarted","Data":"10c30c1d051ff61a9dc843fc1b1b6e5fe4b964576403c5ec523eac4f7b0acbe1"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.129298 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" event={"ID":"151dc193-527a-4506-bcb9-170641b9ea3f","Type":"ContainerStarted","Data":"c393d1d5405f27313978277e9ff4193162615d6d94954f4339b6c095a556ee88"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.146870 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.147207 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-catalog-content\") pod \"community-operators-km59s\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.147234 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-utilities\") pod \"community-operators-km59s\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.147366 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s5fs\" (UniqueName: \"kubernetes.io/projected/2aae37a3-f585-4b9b-94e3-5560e06da84c-kube-api-access-5s5fs\") pod \"community-operators-km59s\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.148369 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.648333554 +0000 UTC m=+121.594701359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.148968 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-catalog-content\") pod \"community-operators-km59s\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.149128 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-utilities\") pod \"community-operators-km59s\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.160400 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56112: no serving certificate available for the kubelet" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.163032 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9fw5r"] Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.164236 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6" event={"ID":"1cd73a43-1689-479b-bae3-6519bdd522b5","Type":"ContainerStarted","Data":"a19dc109b19c5bec770233719de2731bddf0da47e6f87f4c647c02ade64547f6"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.193358 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" event={"ID":"2d469c13-4b04-4c81-9b1b-177250681db4","Type":"ContainerStarted","Data":"e4fdca908e97bdb36358e59b51e9d1204d1e0d4de3ecd1bcb556370dd4eebfeb"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.217256 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" event={"ID":"4f8268de-4dd1-4521-9e5a-b4679996b511","Type":"ContainerStarted","Data":"e288a7943e7642b8a428666629d80a99d08d0ab0a1cb587b24f45965f739cd82"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.226776 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s5fs\" (UniqueName: \"kubernetes.io/projected/2aae37a3-f585-4b9b-94e3-5560e06da84c-kube-api-access-5s5fs\") pod \"community-operators-km59s\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.251981 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" event={"ID":"676cb1dc-3786-44f8-bf27-963da433b3c1","Type":"ContainerStarted","Data":"84c3418158f6f757cdb6ccdcabe2124a9a062a9adf2658e3c3396fcfad356da8"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.252033 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" event={"ID":"676cb1dc-3786-44f8-bf27-963da433b3c1","Type":"ContainerStarted","Data":"698ef6cd4089c90748e863b09feac5f94e0620cef1168809b50bc7f7d91f3785"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.252289 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-km59s" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.253220 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp7h4\" (UniqueName: \"kubernetes.io/projected/014c12c3-e743-45cd-b916-875bf90f7e00-kube-api-access-sp7h4\") pod \"certified-operators-9fw5r\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.253274 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-catalog-content\") pod \"certified-operators-9fw5r\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.253320 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-utilities\") pod \"certified-operators-9fw5r\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.253348 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.253638 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.753626474 +0000 UTC m=+121.699994289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.278753 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" event={"ID":"2c11e87a-9743-40b0-8741-dfec0ce96cb3","Type":"ContainerStarted","Data":"dd02f5451db2f4539023d7b1fea801c0a65da2079ea3d341da9a45ed2843f0fc"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.278828 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" event={"ID":"2c11e87a-9743-40b0-8741-dfec0ce96cb3","Type":"ContainerStarted","Data":"5fd3c0c7fae426fabd70aa7d8cd7f83626c05e6dd85148f02b2fd47bb2eccd03"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.280669 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.293305 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" event={"ID":"b08d05f0-ce9d-49b5-b6ed-f2d02e73b1f4","Type":"ContainerStarted","Data":"b336e9bf1741c6faaf4f87b129384beb7464375b17af8e507a41126bb99c25c5"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.295588 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xtqwz" event={"ID":"3a005177-344e-461b-b8ac-989f18ade02d","Type":"ContainerStarted","Data":"cdbdc94c51ed70d55ae60e16e7c16c5f78ec24ac4fb55120da0021d841e90475"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.296727 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.336738 4936 patch_prober.go:28] interesting pod/console-operator-58897d9998-xtqwz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.336805 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-xtqwz" podUID="3a005177-344e-461b-b8ac-989f18ade02d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.346678 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5nbl4" event={"ID":"921365e5-3846-485e-9afa-270834e81db5","Type":"ContainerStarted","Data":"b57dc9e31feff2972bb7a0152eebf0f84e2cbd07c5c845723143c79eff73d6a9"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.346741 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5nbl4" event={"ID":"921365e5-3846-485e-9afa-270834e81db5","Type":"ContainerStarted","Data":"f72b031e34981cad822ab1545b97f3d2e434de63f0418c5d73849db6690b32cb"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.347645 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.354380 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.354726 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp7h4\" (UniqueName: \"kubernetes.io/projected/014c12c3-e743-45cd-b916-875bf90f7e00-kube-api-access-sp7h4\") pod \"certified-operators-9fw5r\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.354780 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-catalog-content\") pod \"certified-operators-9fw5r\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.354814 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-utilities\") pod \"certified-operators-9fw5r\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.356507 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.856487848 +0000 UTC m=+121.802855663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.356776 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xp5nm"] Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.357856 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-catalog-content\") pod \"certified-operators-9fw5r\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.358019 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-utilities\") pod \"certified-operators-9fw5r\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.359003 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.366456 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" event={"ID":"9027bafb-1a39-4a53-9bcf-778b8859c192","Type":"ContainerStarted","Data":"5e0a5526fe637138ad86076a39fab723933dca03fb4907236829b8475dbe3a3d"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.382527 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qbm7q" event={"ID":"a0abe164-30be-4265-b4b2-4ce8c7b24597","Type":"ContainerStarted","Data":"81889fcf041c34e35f04c7bf3ab3ee3873180163ada12d2f8d8fcfba9ecf6ee4"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.443892 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xp5nm"] Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.458596 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-utilities\") pod \"community-operators-xp5nm\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.458671 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjhn7\" (UniqueName: \"kubernetes.io/projected/a649556e-fd89-4edb-89f7-d1db999c616b-kube-api-access-gjhn7\") pod \"community-operators-xp5nm\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.458717 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.458770 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-catalog-content\") pod \"community-operators-xp5nm\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.460840 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:50.960826303 +0000 UTC m=+121.907194118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.476943 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7jnpn" event={"ID":"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac","Type":"ContainerStarted","Data":"8439beb6463962cf824588bc5506c4c91909786a1e6d317be38e9ac62e58e066"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.476987 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7jnpn" event={"ID":"b65a2c6b-8b39-4742-bd96-6c4cbdc768ac","Type":"ContainerStarted","Data":"ccf03815c205507e44b3371484f1d477ca292ea393d4b937d9b239ece33286ba"} Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.478835 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" gracePeriod=30 Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.494692 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.514193 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp7h4\" (UniqueName: \"kubernetes.io/projected/014c12c3-e743-45cd-b916-875bf90f7e00-kube-api-access-sp7h4\") pod \"certified-operators-9fw5r\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.520324 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l96sz" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.525935 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-m97k8" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.560681 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.561219 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-utilities\") pod \"community-operators-xp5nm\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.561322 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjhn7\" (UniqueName: \"kubernetes.io/projected/a649556e-fd89-4edb-89f7-d1db999c616b-kube-api-access-gjhn7\") pod \"community-operators-xp5nm\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.561449 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-catalog-content\") pod \"community-operators-xp5nm\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.562349 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:51.062322199 +0000 UTC m=+122.008690014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.563531 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-utilities\") pod \"community-operators-xp5nm\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.566186 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-catalog-content\") pod \"community-operators-xp5nm\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.629014 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjhn7\" (UniqueName: \"kubernetes.io/projected/a649556e-fd89-4edb-89f7-d1db999c616b-kube-api-access-gjhn7\") pod \"community-operators-xp5nm\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.663345 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.663778 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:51.163764264 +0000 UTC m=+122.110132079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.703239 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.768942 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.769232 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.769683 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:51.269665982 +0000 UTC m=+122.216033797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.827569 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7jnpn" podStartSLOduration=59.827535905 podStartE2EDuration="59.827535905s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:50.825016226 +0000 UTC m=+121.771384031" watchObservedRunningTime="2026-03-20 16:02:50.827535905 +0000 UTC m=+121.773903720" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.871528 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.871892 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:51.371879048 +0000 UTC m=+122.318246853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.880612 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-xlnmq" podStartSLOduration=59.880593936 podStartE2EDuration="59.880593936s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:50.866650245 +0000 UTC m=+121.813018060" watchObservedRunningTime="2026-03-20 16:02:50.880593936 +0000 UTC m=+121.826961751" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.883625 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.888737 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:50 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:50 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:50 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.888823 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.900393 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" podStartSLOduration=59.900370207 podStartE2EDuration="59.900370207s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:50.899336419 +0000 UTC m=+121.845704254" watchObservedRunningTime="2026-03-20 16:02:50.900370207 +0000 UTC m=+121.846738022" Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.973268 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:50 crc kubenswrapper[4936]: E0320 16:02:50.973651 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:51.473633322 +0000 UTC m=+122.420001137 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:50 crc kubenswrapper[4936]: I0320 16:02:50.983097 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" podStartSLOduration=59.98307578 podStartE2EDuration="59.98307578s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:50.982102143 +0000 UTC m=+121.928469958" watchObservedRunningTime="2026-03-20 16:02:50.98307578 +0000 UTC m=+121.929443595" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.084274 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.084713 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:51.58470001 +0000 UTC m=+122.531067825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.126725 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" podStartSLOduration=60.126702149 podStartE2EDuration="1m0.126702149s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.086436308 +0000 UTC m=+122.032804123" watchObservedRunningTime="2026-03-20 16:02:51.126702149 +0000 UTC m=+122.073069974" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.162742 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-58dfj" podStartSLOduration=60.162718204 podStartE2EDuration="1m0.162718204s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.160910925 +0000 UTC m=+122.107278740" watchObservedRunningTime="2026-03-20 16:02:51.162718204 +0000 UTC m=+122.109086019" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.190137 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.190489 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:51.690473054 +0000 UTC m=+122.636840869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.198631 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5cpcn"] Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.255052 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5l6qv" podStartSLOduration=60.25503266 podStartE2EDuration="1m0.25503266s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.253892899 +0000 UTC m=+122.200260714" watchObservedRunningTime="2026-03-20 16:02:51.25503266 +0000 UTC m=+122.201400475" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.301232 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.301789 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:51.801757848 +0000 UTC m=+122.748125653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.332297 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tvs2g" podStartSLOduration=60.332269933 podStartE2EDuration="1m0.332269933s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.286619154 +0000 UTC m=+122.232986979" watchObservedRunningTime="2026-03-20 16:02:51.332269933 +0000 UTC m=+122.278637748" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.345629 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4p8zd"] Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.345944 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" podUID="20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" containerName="controller-manager" containerID="cri-o://bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca" gracePeriod=30 Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.377934 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl"] Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.415295 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-km59s"] Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.425949 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-j54jw" podStartSLOduration=60.425928325 podStartE2EDuration="1m0.425928325s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.370885209 +0000 UTC m=+122.317253044" watchObservedRunningTime="2026-03-20 16:02:51.425928325 +0000 UTC m=+122.372296140" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.440091 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.440519 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:51.940502154 +0000 UTC m=+122.886869959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.468565 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qxmr7" podStartSLOduration=60.468532191 podStartE2EDuration="1m0.468532191s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.467158993 +0000 UTC m=+122.413526808" watchObservedRunningTime="2026-03-20 16:02:51.468532191 +0000 UTC m=+122.414900006" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.522602 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-42wg6" podStartSLOduration=60.522576139 podStartE2EDuration="1m0.522576139s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.520268936 +0000 UTC m=+122.466636741" watchObservedRunningTime="2026-03-20 16:02:51.522576139 +0000 UTC m=+122.468943954" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.530352 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56124: no serving certificate available for the kubelet" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.550487 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.554653 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.054627826 +0000 UTC m=+123.000995661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.609939 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5cpcn" event={"ID":"7964eb91-819c-4853-a598-9b93346f496f","Type":"ContainerStarted","Data":"b44f630a17ca56a75599441591c25625a11f2ceaa1a320105c57b3e94b80ba8e"} Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.628481 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km59s" event={"ID":"2aae37a3-f585-4b9b-94e3-5560e06da84c","Type":"ContainerStarted","Data":"ebd485fa20d242621384574fafa29831784c0ffa71f3c5fcd346fec2e83660a9"} Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.642401 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xp5nm"] Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.651830 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-xtqwz" podStartSLOduration=60.651796664 podStartE2EDuration="1m0.651796664s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.637290027 +0000 UTC m=+122.583657842" watchObservedRunningTime="2026-03-20 16:02:51.651796664 +0000 UTC m=+122.598164469" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.666364 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vm2rn" podStartSLOduration=60.666344562 podStartE2EDuration="1m0.666344562s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.664036419 +0000 UTC m=+122.610404244" watchObservedRunningTime="2026-03-20 16:02:51.666344562 +0000 UTC m=+122.612712377" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.673307 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.675628 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.175591935 +0000 UTC m=+123.121959750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.776502 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.776917 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.276902327 +0000 UTC m=+123.223270142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.782914 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dbjm7" podStartSLOduration=60.782898041 podStartE2EDuration="1m0.782898041s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.780017262 +0000 UTC m=+122.726385077" watchObservedRunningTime="2026-03-20 16:02:51.782898041 +0000 UTC m=+122.729265846" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.822313 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-xtqwz" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.847122 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9fw5r"] Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.854722 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:51 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:51 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:51 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.854784 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.878973 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.879249 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.379229145 +0000 UTC m=+123.325596950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.880661 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.881188 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.381179398 +0000 UTC m=+123.327547213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.900401 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-5nbl4" podStartSLOduration=10.900383994 podStartE2EDuration="10.900383994s" podCreationTimestamp="2026-03-20 16:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.894354629 +0000 UTC m=+122.840722444" watchObservedRunningTime="2026-03-20 16:02:51.900383994 +0000 UTC m=+122.846751809" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.915117 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-njqd9"] Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.916652 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:51 crc kubenswrapper[4936]: W0320 16:02:51.926996 4936 reflector.go:561] object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb": failed to list *v1.Secret: secrets "redhat-marketplace-dockercfg-x2ctb" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.931955 4936 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-x2ctb\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"redhat-marketplace-dockercfg-x2ctb\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.937991 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-njqd9"] Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.951745 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=1.951726538 podStartE2EDuration="1.951726538s" podCreationTimestamp="2026-03-20 16:02:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:51.950920326 +0000 UTC m=+122.897288131" watchObservedRunningTime="2026-03-20 16:02:51.951726538 +0000 UTC m=+122.898094353" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.982324 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.982708 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj5v4\" (UniqueName: \"kubernetes.io/projected/959c1ced-7501-4754-8e17-c3a420edd1b8-kube-api-access-dj5v4\") pod \"redhat-marketplace-njqd9\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:51 crc kubenswrapper[4936]: E0320 16:02:51.982821 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.482773358 +0000 UTC m=+123.429141173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.983073 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-utilities\") pod \"redhat-marketplace-njqd9\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:51 crc kubenswrapper[4936]: I0320 16:02:51.983500 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-catalog-content\") pod \"redhat-marketplace-njqd9\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.085365 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-utilities\") pod \"redhat-marketplace-njqd9\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.085415 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-catalog-content\") pod \"redhat-marketplace-njqd9\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.085495 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj5v4\" (UniqueName: \"kubernetes.io/projected/959c1ced-7501-4754-8e17-c3a420edd1b8-kube-api-access-dj5v4\") pod \"redhat-marketplace-njqd9\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.085530 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.085921 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.585905229 +0000 UTC m=+123.532273044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.086134 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-catalog-content\") pod \"redhat-marketplace-njqd9\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.086386 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-utilities\") pod \"redhat-marketplace-njqd9\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.146958 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj5v4\" (UniqueName: \"kubernetes.io/projected/959c1ced-7501-4754-8e17-c3a420edd1b8-kube-api-access-dj5v4\") pod \"redhat-marketplace-njqd9\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.175926 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.186348 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.186444 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-serving-cert\") pod \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.186481 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-config\") pod \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.186551 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-proxy-ca-bundles\") pod \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.186573 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtxgl\" (UniqueName: \"kubernetes.io/projected/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-kube-api-access-qtxgl\") pod \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.186633 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-client-ca\") pod \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\" (UID: \"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21\") " Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.188076 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-config" (OuterVolumeSpecName: "config") pod "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" (UID: "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.188230 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.688206998 +0000 UTC m=+123.634574813 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.189639 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" (UID: "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.191735 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-client-ca" (OuterVolumeSpecName: "client-ca") pod "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" (UID: "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.203802 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" (UID: "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.206819 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-kube-api-access-qtxgl" (OuterVolumeSpecName: "kube-api-access-qtxgl") pod "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" (UID: "20cd0c30-bf76-4164-b1b8-2c0de4bdbb21"). InnerVolumeSpecName "kube-api-access-qtxgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.288712 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.289242 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtxgl\" (UniqueName: \"kubernetes.io/projected/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-kube-api-access-qtxgl\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.289257 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.289268 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.289278 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.289287 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.289742 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.789724335 +0000 UTC m=+123.736092150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.327388 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pqjxj"] Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.327638 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" containerName="controller-manager" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.327654 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" containerName="controller-manager" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.327766 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" containerName="controller-manager" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.331822 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.332554 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqjxj"] Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.389930 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.390053 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-catalog-content\") pod \"redhat-marketplace-pqjxj\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.390169 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.890132942 +0000 UTC m=+123.836500757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.390401 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.390461 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rwpr\" (UniqueName: \"kubernetes.io/projected/cd71509d-889b-498c-b5ec-55e6902b08f7-kube-api-access-5rwpr\") pod \"redhat-marketplace-pqjxj\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.390496 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-utilities\") pod \"redhat-marketplace-pqjxj\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.390857 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.890842121 +0000 UTC m=+123.837209936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.492320 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.492582 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.992534203 +0000 UTC m=+123.938902038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.492759 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.492814 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rwpr\" (UniqueName: \"kubernetes.io/projected/cd71509d-889b-498c-b5ec-55e6902b08f7-kube-api-access-5rwpr\") pod \"redhat-marketplace-pqjxj\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.492847 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-utilities\") pod \"redhat-marketplace-pqjxj\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.492964 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-catalog-content\") pod \"redhat-marketplace-pqjxj\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.493152 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:52.99312742 +0000 UTC m=+123.939495415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.493506 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-utilities\") pod \"redhat-marketplace-pqjxj\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.493625 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-catalog-content\") pod \"redhat-marketplace-pqjxj\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.513028 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rwpr\" (UniqueName: \"kubernetes.io/projected/cd71509d-889b-498c-b5ec-55e6902b08f7-kube-api-access-5rwpr\") pod \"redhat-marketplace-pqjxj\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.531063 4936 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.593427 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.593646 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:53.093614089 +0000 UTC m=+124.039981904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.593892 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.594183 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:53.094170394 +0000 UTC m=+124.040538209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.636651 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xp5nm" event={"ID":"a649556e-fd89-4edb-89f7-d1db999c616b","Type":"ContainerDied","Data":"c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.636605 4936 generic.go:334] "Generic (PLEG): container finished" podID="a649556e-fd89-4edb-89f7-d1db999c616b" containerID="c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156" exitCode=0 Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.636920 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xp5nm" event={"ID":"a649556e-fd89-4edb-89f7-d1db999c616b","Type":"ContainerStarted","Data":"c475781c82d1343e7b359032f9f5d3316daf4be1f437270b6f129d1fd072b7fc"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.640496 4936 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.647706 4936 generic.go:334] "Generic (PLEG): container finished" podID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerID="74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22" exitCode=0 Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.647857 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km59s" event={"ID":"2aae37a3-f585-4b9b-94e3-5560e06da84c","Type":"ContainerDied","Data":"74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.664919 4936 generic.go:334] "Generic (PLEG): container finished" podID="e84e4377-e478-4c12-85bf-b236885152b2" containerID="110680cf2e9247bb1db2afd5997a349c5a6285e2bf3818bb3fa35e00d004f836" exitCode=0 Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.665011 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" event={"ID":"e84e4377-e478-4c12-85bf-b236885152b2","Type":"ContainerDied","Data":"110680cf2e9247bb1db2afd5997a349c5a6285e2bf3818bb3fa35e00d004f836"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.672408 4936 generic.go:334] "Generic (PLEG): container finished" podID="7964eb91-819c-4853-a598-9b93346f496f" containerID="69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5" exitCode=0 Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.673309 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5cpcn" event={"ID":"7964eb91-819c-4853-a598-9b93346f496f","Type":"ContainerDied","Data":"69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.685312 4936 generic.go:334] "Generic (PLEG): container finished" podID="014c12c3-e743-45cd-b916-875bf90f7e00" containerID="109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88" exitCode=0 Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.685660 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fw5r" event={"ID":"014c12c3-e743-45cd-b916-875bf90f7e00","Type":"ContainerDied","Data":"109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.685709 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fw5r" event={"ID":"014c12c3-e743-45cd-b916-875bf90f7e00","Type":"ContainerStarted","Data":"4cf01fd44fb32e1558e4fee63b79fbd849f8c89d019d20112213cf946203ea74"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.691829 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" event={"ID":"2d469c13-4b04-4c81-9b1b-177250681db4","Type":"ContainerStarted","Data":"3f6ff54c8e37a47b521c9141d1ee6f6ffa45c83b49af127a10e664d27285f0cc"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.691880 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" event={"ID":"2d469c13-4b04-4c81-9b1b-177250681db4","Type":"ContainerStarted","Data":"dbae139cfb9fe793957dbcc8c7591027847d45b49177cd52b09d31fbb1d0ca11"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.694756 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.695518 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:53.195502656 +0000 UTC m=+124.141870471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.696525 4936 generic.go:334] "Generic (PLEG): container finished" podID="20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" containerID="bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca" exitCode=0 Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.696877 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" podUID="d3739ff0-d3fa-4243-9529-209e3ec9e4e4" containerName="route-controller-manager" containerID="cri-o://58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029" gracePeriod=30 Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.696620 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.696567 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" event={"ID":"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21","Type":"ContainerDied","Data":"bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.697786 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" event={"ID":"20cd0c30-bf76-4164-b1b8-2c0de4bdbb21","Type":"ContainerDied","Data":"91b9503fdd4ee9b28c1eae0da7d03d7a07f8ba84c60c9cc52a859fbbb820bc82"} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.697816 4936 scope.go:117] "RemoveContainer" containerID="bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.727826 4936 scope.go:117] "RemoveContainer" containerID="bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.729285 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca\": container with ID starting with bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca not found: ID does not exist" containerID="bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.729334 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca"} err="failed to get container status \"bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca\": rpc error: code = NotFound desc = could not find container \"bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca\": container with ID starting with bb797c278362db071eab9cd2d6575268f6f2c291f1432522ae245f1ff46c0cca not found: ID does not exist" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.753271 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4p8zd"] Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.755751 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4p8zd"] Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.796863 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.797875 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:53.297857166 +0000 UTC m=+124.244224981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.854305 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:52 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:52 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:52 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.854384 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.897207 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kg66h"] Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.897637 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.897801 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 16:02:53.397767699 +0000 UTC m=+124.344135514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.898048 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:52 crc kubenswrapper[4936]: E0320 16:02:52.898415 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 16:02:53.398402627 +0000 UTC m=+124.344770442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8tlt4" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.898852 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.902158 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.911780 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kg66h"] Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.955899 4936 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4p8zd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: i/o timeout" start-of-body= Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.955988 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4p8zd" podUID="20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: i/o timeout" Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.984831 4936 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-20T16:02:52.531113799Z","Handler":null,"Name":""} Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.990165 4936 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 20 16:02:52 crc kubenswrapper[4936]: I0320 16:02:52.990198 4936 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.003066 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.003429 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-utilities\") pod \"redhat-operators-kg66h\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.003463 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhdlt\" (UniqueName: \"kubernetes.io/projected/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-kube-api-access-nhdlt\") pod \"redhat-operators-kg66h\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.003503 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-catalog-content\") pod \"redhat-operators-kg66h\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.009294 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.052583 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.053852 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.056084 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.056989 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.058660 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.081041 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.104276 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-serving-cert\") pod \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.104398 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-config\") pod \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.104448 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgp2g\" (UniqueName: \"kubernetes.io/projected/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-kube-api-access-zgp2g\") pod \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.104481 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-client-ca\") pod \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\" (UID: \"d3739ff0-d3fa-4243-9529-209e3ec9e4e4\") " Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.104690 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhdlt\" (UniqueName: \"kubernetes.io/projected/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-kube-api-access-nhdlt\") pod \"redhat-operators-kg66h\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.104727 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.104753 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.104777 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-catalog-content\") pod \"redhat-operators-kg66h\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.104987 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.105045 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-utilities\") pod \"redhat-operators-kg66h\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.105798 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-config" (OuterVolumeSpecName: "config") pod "d3739ff0-d3fa-4243-9529-209e3ec9e4e4" (UID: "d3739ff0-d3fa-4243-9529-209e3ec9e4e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.105942 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-catalog-content\") pod \"redhat-operators-kg66h\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.105975 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-utilities\") pod \"redhat-operators-kg66h\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.106326 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-client-ca" (OuterVolumeSpecName: "client-ca") pod "d3739ff0-d3fa-4243-9529-209e3ec9e4e4" (UID: "d3739ff0-d3fa-4243-9529-209e3ec9e4e4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.110535 4936 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.110608 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.113927 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-kube-api-access-zgp2g" (OuterVolumeSpecName: "kube-api-access-zgp2g") pod "d3739ff0-d3fa-4243-9529-209e3ec9e4e4" (UID: "d3739ff0-d3fa-4243-9529-209e3ec9e4e4"). InnerVolumeSpecName "kube-api-access-zgp2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.121861 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhdlt\" (UniqueName: \"kubernetes.io/projected/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-kube-api-access-nhdlt\") pod \"redhat-operators-kg66h\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.130104 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d3739ff0-d3fa-4243-9529-209e3ec9e4e4" (UID: "d3739ff0-d3fa-4243-9529-209e3ec9e4e4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.155175 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8tlt4\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.206796 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.206863 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.206946 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.207067 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.207089 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgp2g\" (UniqueName: \"kubernetes.io/projected/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-kube-api-access-zgp2g\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.207102 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.207114 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3739ff0-d3fa-4243-9529-209e3ec9e4e4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.218619 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.224706 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.264227 4936 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/redhat-marketplace-njqd9" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.264351 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.294506 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6xpds"] Mar 20 16:02:53 crc kubenswrapper[4936]: E0320 16:02:53.294795 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3739ff0-d3fa-4243-9529-209e3ec9e4e4" containerName="route-controller-manager" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.294810 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3739ff0-d3fa-4243-9529-209e3ec9e4e4" containerName="route-controller-manager" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.294932 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3739ff0-d3fa-4243-9529-209e3ec9e4e4" containerName="route-controller-manager" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.296115 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.306301 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.308524 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlwzp\" (UniqueName: \"kubernetes.io/projected/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-kube-api-access-qlwzp\") pod \"redhat-operators-6xpds\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.308609 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-utilities\") pod \"redhat-operators-6xpds\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.308681 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-catalog-content\") pod \"redhat-operators-6xpds\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.314518 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6xpds"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.314851 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.329207 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.330007 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.376272 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.377376 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.381494 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-559b965cdc-6sctl"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.392577 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.392695 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.393443 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-559b965cdc-6sctl"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.396565 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.397744 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.398051 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.398598 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.398746 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.398849 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.398889 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.406745 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409510 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-proxy-ca-bundles\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409566 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-catalog-content\") pod \"redhat-operators-6xpds\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409590 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-config\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409623 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-config\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409655 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-client-ca\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409681 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlwzp\" (UniqueName: \"kubernetes.io/projected/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-kube-api-access-qlwzp\") pod \"redhat-operators-6xpds\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409699 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-client-ca\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409726 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8phk\" (UniqueName: \"kubernetes.io/projected/ff75dce3-7056-4983-99d8-f67217bdba2f-kube-api-access-k8phk\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409754 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmh2z\" (UniqueName: \"kubernetes.io/projected/67ff8a83-b028-4b8f-9fee-58847d147c08-kube-api-access-zmh2z\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409790 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff75dce3-7056-4983-99d8-f67217bdba2f-serving-cert\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409806 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ff8a83-b028-4b8f-9fee-58847d147c08-serving-cert\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.409823 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-utilities\") pod \"redhat-operators-6xpds\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.410216 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-utilities\") pod \"redhat-operators-6xpds\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.410472 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-catalog-content\") pod \"redhat-operators-6xpds\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.438979 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlwzp\" (UniqueName: \"kubernetes.io/projected/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-kube-api-access-qlwzp\") pod \"redhat-operators-6xpds\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.518742 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-client-ca\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.519226 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8phk\" (UniqueName: \"kubernetes.io/projected/ff75dce3-7056-4983-99d8-f67217bdba2f-kube-api-access-k8phk\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.519269 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmh2z\" (UniqueName: \"kubernetes.io/projected/67ff8a83-b028-4b8f-9fee-58847d147c08-kube-api-access-zmh2z\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.519326 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff75dce3-7056-4983-99d8-f67217bdba2f-serving-cert\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.519351 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ff8a83-b028-4b8f-9fee-58847d147c08-serving-cert\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.519398 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-proxy-ca-bundles\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.519434 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-config\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.519690 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-config\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.519730 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-client-ca\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.522063 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-client-ca\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.523340 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-client-ca\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.523794 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-config\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.523917 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-config\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.526794 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-proxy-ca-bundles\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.529066 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ff8a83-b028-4b8f-9fee-58847d147c08-serving-cert\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.557436 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kg66h"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.559197 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff75dce3-7056-4983-99d8-f67217bdba2f-serving-cert\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.568095 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmh2z\" (UniqueName: \"kubernetes.io/projected/67ff8a83-b028-4b8f-9fee-58847d147c08-kube-api-access-zmh2z\") pod \"route-controller-manager-5989f75564-7hm4h\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.570514 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8phk\" (UniqueName: \"kubernetes.io/projected/ff75dce3-7056-4983-99d8-f67217bdba2f-kube-api-access-k8phk\") pod \"controller-manager-559b965cdc-6sctl\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.612527 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.620212 4936 patch_prober.go:28] interesting pod/downloads-7954f5f757-dzgl2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.620242 4936 patch_prober.go:28] interesting pod/downloads-7954f5f757-dzgl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.620276 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-dzgl2" podUID="bf0f9e97-bddb-4054-99dd-d10428233e9f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.620316 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dzgl2" podUID="bf0f9e97-bddb-4054-99dd-d10428233e9f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.703277 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.711494 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.742715 4936 generic.go:334] "Generic (PLEG): container finished" podID="d3739ff0-d3fa-4243-9529-209e3ec9e4e4" containerID="58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029" exitCode=0 Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.742800 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" event={"ID":"d3739ff0-d3fa-4243-9529-209e3ec9e4e4","Type":"ContainerDied","Data":"58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029"} Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.742837 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" event={"ID":"d3739ff0-d3fa-4243-9529-209e3ec9e4e4","Type":"ContainerDied","Data":"4f5a5b84bc720f1b3933512c86d90d7845c5ba5340b66137bd4cf5e4ca7d4cec"} Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.742859 4936 scope.go:117] "RemoveContainer" containerID="58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.743008 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.753753 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg66h" event={"ID":"b7e7487c-e0c4-47a9-9875-b1c41e14eecc","Type":"ContainerStarted","Data":"f4d3633c8d428598577285ffd783dca5141f6a3ad903f6323e5ebb24fb49022c"} Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.813086 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" event={"ID":"2d469c13-4b04-4c81-9b1b-177250681db4","Type":"ContainerStarted","Data":"5e9b4466035d9ba8a8e94fe88323a950525032ccd58b70805575af0a5a4bc743"} Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.847492 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.847572 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.847623 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.850705 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.850730 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.851881 4936 scope.go:117] "RemoveContainer" containerID="58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.856925 4936 scope.go:117] "RemoveContainer" containerID="034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.859756 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:53 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:53 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:53 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.859828 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:53 crc kubenswrapper[4936]: E0320 16:02:53.869261 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029\": container with ID starting with 58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029 not found: ID does not exist" containerID="58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.869409 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029"} err="failed to get container status \"58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029\": rpc error: code = NotFound desc = could not find container \"58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029\": container with ID starting with 58675b87773b7b3a25a957a15020068065960397290629fca37f774fe84f8029 not found: ID does not exist" Mar 20 16:02:53 crc kubenswrapper[4936]: E0320 16:02:53.895060 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.916445 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20cd0c30-bf76-4164-b1b8-2c0de4bdbb21" path="/var/lib/kubelet/pods/20cd0c30-bf76-4164-b1b8-2c0de4bdbb21/volumes" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.917482 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.918186 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cf7gl"] Mar 20 16:02:53 crc kubenswrapper[4936]: E0320 16:02:53.936195 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.937391 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.938279 4936 patch_prober.go:28] interesting pod/apiserver-76f77b778f-jgfjb container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]log ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]etcd ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/generic-apiserver-start-informers ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/max-in-flight-filter ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 20 16:02:53 crc kubenswrapper[4936]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/project.openshift.io-projectcache ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/openshift.io-startinformers ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 20 16:02:53 crc kubenswrapper[4936]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 20 16:02:53 crc kubenswrapper[4936]: livez check failed Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.938349 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" podUID="151dc193-527a-4506-bcb9-170641b9ea3f" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.951784 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.952204 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.966864 4936 patch_prober.go:28] interesting pod/console-f9d7485db-6q5fl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.966944 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6q5fl" podUID="c52c0329-ca68-4d95-8341-ab8191c3a186" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 20 16:02:53 crc kubenswrapper[4936]: I0320 16:02:53.975461 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-dgtvg" podStartSLOduration=12.975441761 podStartE2EDuration="12.975441761s" podCreationTimestamp="2026-03-20 16:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:53.915933793 +0000 UTC m=+124.862301608" watchObservedRunningTime="2026-03-20 16:02:53.975441761 +0000 UTC m=+124.921809566" Mar 20 16:02:53 crc kubenswrapper[4936]: E0320 16:02:53.996747 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:02:53 crc kubenswrapper[4936]: E0320 16:02:53.996846 4936 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerName="kube-multus-additional-cni-plugins" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.042232 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqjxj"] Mar 20 16:02:54 crc kubenswrapper[4936]: W0320 16:02:54.089481 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53d1ed6e_605c_46b9_b053_ba5a899db885.slice/crio-39e74f84a3982bf3cb8fadb8bc8e4cdf799df704fd09e84c35e6a30448ba9d57 WatchSource:0}: Error finding container 39e74f84a3982bf3cb8fadb8bc8e4cdf799df704fd09e84c35e6a30448ba9d57: Status 404 returned error can't find the container with id 39e74f84a3982bf3cb8fadb8bc8e4cdf799df704fd09e84c35e6a30448ba9d57 Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.101436 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8tlt4"] Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.108308 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-njqd9"] Mar 20 16:02:54 crc kubenswrapper[4936]: W0320 16:02:54.125737 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod959c1ced_7501_4754_8e17_c3a420edd1b8.slice/crio-98f458d1fd0de63f1e737aaaadb24d49b86604459029629917df309d5134188d WatchSource:0}: Error finding container 98f458d1fd0de63f1e737aaaadb24d49b86604459029629917df309d5134188d: Status 404 returned error can't find the container with id 98f458d1fd0de63f1e737aaaadb24d49b86604459029629917df309d5134188d Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.138996 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56126: no serving certificate available for the kubelet" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.140538 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.312354 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6xpds"] Mar 20 16:02:54 crc kubenswrapper[4936]: W0320 16:02:54.340043 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b3ba1a3_fe09_48fe_a584_3cbd20c2225a.slice/crio-5901809f687d4eb42800894c2d8bd5fe947ad4f155d60efbfce75e4ba4b72c70 WatchSource:0}: Error finding container 5901809f687d4eb42800894c2d8bd5fe947ad4f155d60efbfce75e4ba4b72c70: Status 404 returned error can't find the container with id 5901809f687d4eb42800894c2d8bd5fe947ad4f155d60efbfce75e4ba4b72c70 Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.393437 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56130: no serving certificate available for the kubelet" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.417602 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.527174 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h"] Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.541164 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-559b965cdc-6sctl"] Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.561486 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84e4377-e478-4c12-85bf-b236885152b2-secret-volume\") pod \"e84e4377-e478-4c12-85bf-b236885152b2\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.561659 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84e4377-e478-4c12-85bf-b236885152b2-config-volume\") pod \"e84e4377-e478-4c12-85bf-b236885152b2\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.561694 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhp77\" (UniqueName: \"kubernetes.io/projected/e84e4377-e478-4c12-85bf-b236885152b2-kube-api-access-rhp77\") pod \"e84e4377-e478-4c12-85bf-b236885152b2\" (UID: \"e84e4377-e478-4c12-85bf-b236885152b2\") " Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.564676 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e84e4377-e478-4c12-85bf-b236885152b2-config-volume" (OuterVolumeSpecName: "config-volume") pod "e84e4377-e478-4c12-85bf-b236885152b2" (UID: "e84e4377-e478-4c12-85bf-b236885152b2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.570498 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e84e4377-e478-4c12-85bf-b236885152b2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e84e4377-e478-4c12-85bf-b236885152b2" (UID: "e84e4377-e478-4c12-85bf-b236885152b2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.572162 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e84e4377-e478-4c12-85bf-b236885152b2-kube-api-access-rhp77" (OuterVolumeSpecName: "kube-api-access-rhp77") pod "e84e4377-e478-4c12-85bf-b236885152b2" (UID: "e84e4377-e478-4c12-85bf-b236885152b2"). InnerVolumeSpecName "kube-api-access-rhp77". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.666857 4936 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84e4377-e478-4c12-85bf-b236885152b2-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.666883 4936 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84e4377-e478-4c12-85bf-b236885152b2-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.666894 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhp77\" (UniqueName: \"kubernetes.io/projected/e84e4377-e478-4c12-85bf-b236885152b2-kube-api-access-rhp77\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.849936 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.853347 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:54 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:54 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:54 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.853391 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.906521 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" event={"ID":"53d1ed6e-605c-46b9-b053-ba5a899db885","Type":"ContainerStarted","Data":"ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14"} Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.906614 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" event={"ID":"53d1ed6e-605c-46b9-b053-ba5a899db885","Type":"ContainerStarted","Data":"39e74f84a3982bf3cb8fadb8bc8e4cdf799df704fd09e84c35e6a30448ba9d57"} Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.906673 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.941162 4936 generic.go:334] "Generic (PLEG): container finished" podID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerID="61c8748b50301280d495ee41d29e10d6a82c7d901f6e4fcb8d45d45679fd0ffc" exitCode=0 Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.941232 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xpds" event={"ID":"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a","Type":"ContainerDied","Data":"61c8748b50301280d495ee41d29e10d6a82c7d901f6e4fcb8d45d45679fd0ffc"} Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.941259 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xpds" event={"ID":"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a","Type":"ContainerStarted","Data":"5901809f687d4eb42800894c2d8bd5fe947ad4f155d60efbfce75e4ba4b72c70"} Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.956034 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.956288 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567040-9g8q5" event={"ID":"e84e4377-e478-4c12-85bf-b236885152b2","Type":"ContainerDied","Data":"e361365e6b1edbc6ddd5ab373f6dc57b182135e658f87174c7167d0ee85d4090"} Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.956343 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e361365e6b1edbc6ddd5ab373f6dc57b182135e658f87174c7167d0ee85d4090" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.961487 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" podStartSLOduration=63.961473876 podStartE2EDuration="1m3.961473876s" podCreationTimestamp="2026-03-20 16:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:54.931535267 +0000 UTC m=+125.877903102" watchObservedRunningTime="2026-03-20 16:02:54.961473876 +0000 UTC m=+125.907841691" Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.963304 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc98fe2f-d705-4a0a-83e2-0f5798eaf509","Type":"ContainerStarted","Data":"9d58262437b86bc7d7a0016012869f0a465ce2352e063b5c5bcb363bb30b9792"} Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.976479 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" event={"ID":"67ff8a83-b028-4b8f-9fee-58847d147c08","Type":"ContainerStarted","Data":"c490d4a14c8c59730e495ce35969b205b7349bdd344ba7b380605935baf7b12c"} Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.988960 4936 generic.go:334] "Generic (PLEG): container finished" podID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerID="2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2" exitCode=0 Mar 20 16:02:54 crc kubenswrapper[4936]: I0320 16:02:54.989130 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg66h" event={"ID":"b7e7487c-e0c4-47a9-9875-b1c41e14eecc","Type":"ContainerDied","Data":"2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2"} Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.000272 4936 generic.go:334] "Generic (PLEG): container finished" podID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerID="e22926ace22dfafb0ec423692d3f189059d67ddb67183f81e7c57d62e858ac98" exitCode=0 Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.000406 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqjxj" event={"ID":"cd71509d-889b-498c-b5ec-55e6902b08f7","Type":"ContainerDied","Data":"e22926ace22dfafb0ec423692d3f189059d67ddb67183f81e7c57d62e858ac98"} Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.000478 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqjxj" event={"ID":"cd71509d-889b-498c-b5ec-55e6902b08f7","Type":"ContainerStarted","Data":"f7293e3ed6f67cd098b73da8f73f22f6eb1b2c3843a0d8516a747b2d94b31f81"} Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.067983 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.077463 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62"} Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.078556 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.092474 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" event={"ID":"ff75dce3-7056-4983-99d8-f67217bdba2f","Type":"ContainerStarted","Data":"d51847f92beee533240ab9a08d9020b49d4d6b470ba05e406e47b2aa4fcc5438"} Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.093870 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.102727 4936 patch_prober.go:28] interesting pod/controller-manager-559b965cdc-6sctl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" start-of-body= Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.102808 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" podUID="ff75dce3-7056-4983-99d8-f67217bdba2f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.103808 4936 generic.go:334] "Generic (PLEG): container finished" podID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerID="ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36" exitCode=0 Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.104611 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njqd9" event={"ID":"959c1ced-7501-4754-8e17-c3a420edd1b8","Type":"ContainerDied","Data":"ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36"} Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.104645 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njqd9" event={"ID":"959c1ced-7501-4754-8e17-c3a420edd1b8","Type":"ContainerStarted","Data":"98f458d1fd0de63f1e737aaaadb24d49b86604459029629917df309d5134188d"} Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.118565 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.118507922 podStartE2EDuration="17.118507922s" podCreationTimestamp="2026-03-20 16:02:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:55.103189713 +0000 UTC m=+126.049557548" watchObservedRunningTime="2026-03-20 16:02:55.118507922 +0000 UTC m=+126.064875737" Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.120905 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-57jx7" Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.145905 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" podStartSLOduration=3.145875761 podStartE2EDuration="3.145875761s" podCreationTimestamp="2026-03-20 16:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:55.145219623 +0000 UTC m=+126.091587438" watchObservedRunningTime="2026-03-20 16:02:55.145875761 +0000 UTC m=+126.092243576" Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.857274 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:55 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:55 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:55 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.857630 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:55 crc kubenswrapper[4936]: I0320 16:02:55.873321 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3739ff0-d3fa-4243-9529-209e3ec9e4e4" path="/var/lib/kubelet/pods/d3739ff0-d3fa-4243-9529-209e3ec9e4e4/volumes" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.013469 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 16:02:56 crc kubenswrapper[4936]: E0320 16:02:56.013869 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e84e4377-e478-4c12-85bf-b236885152b2" containerName="collect-profiles" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.013890 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="e84e4377-e478-4c12-85bf-b236885152b2" containerName="collect-profiles" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.018230 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="e84e4377-e478-4c12-85bf-b236885152b2" containerName="collect-profiles" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.020941 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.021080 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.026757 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.033463 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.142561 4936 generic.go:334] "Generic (PLEG): container finished" podID="bc98fe2f-d705-4a0a-83e2-0f5798eaf509" containerID="753b14a1009465dbac1f1eae52d6aade4bec1575cc7c0506115e2f1e8cd16dc8" exitCode=0 Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.142643 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc98fe2f-d705-4a0a-83e2-0f5798eaf509","Type":"ContainerDied","Data":"753b14a1009465dbac1f1eae52d6aade4bec1575cc7c0506115e2f1e8cd16dc8"} Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.148570 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" event={"ID":"67ff8a83-b028-4b8f-9fee-58847d147c08","Type":"ContainerStarted","Data":"dd242b81c4774344d9415e73f518ad3898125c58bb421ddbcac6f967fd2bea6e"} Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.149561 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.154596 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.157095 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" event={"ID":"ff75dce3-7056-4983-99d8-f67217bdba2f","Type":"ContainerStarted","Data":"f56f07ae6c26c400f7195662e3f6cd991836ba2513903775ffed98ca15c86937"} Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.164017 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.184424 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" podStartSLOduration=4.184153944 podStartE2EDuration="4.184153944s" podCreationTimestamp="2026-03-20 16:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:56.180398642 +0000 UTC m=+127.126766457" watchObservedRunningTime="2026-03-20 16:02:56.184153944 +0000 UTC m=+127.130521759" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.193927 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.194031 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.295904 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.296142 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.298871 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.317731 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.356481 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.912190 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:56 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:56 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:56 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.912888 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:56 crc kubenswrapper[4936]: I0320 16:02:56.931324 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 16:02:56 crc kubenswrapper[4936]: W0320 16:02:56.959135 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod23ed9f1c_439a_4626_a8e3_5c804cfd6bcd.slice/crio-a823287493fbfa065447b454e3b60a18d335fbfc6c4dac2f8b2f0cedf4443d54 WatchSource:0}: Error finding container a823287493fbfa065447b454e3b60a18d335fbfc6c4dac2f8b2f0cedf4443d54: Status 404 returned error can't find the container with id a823287493fbfa065447b454e3b60a18d335fbfc6c4dac2f8b2f0cedf4443d54 Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.224102 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd","Type":"ContainerStarted","Data":"a823287493fbfa065447b454e3b60a18d335fbfc6c4dac2f8b2f0cedf4443d54"} Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.581672 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.727703 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kube-api-access\") pod \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\" (UID: \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\") " Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.730778 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kubelet-dir\") pod \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\" (UID: \"bc98fe2f-d705-4a0a-83e2-0f5798eaf509\") " Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.731034 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bc98fe2f-d705-4a0a-83e2-0f5798eaf509" (UID: "bc98fe2f-d705-4a0a-83e2-0f5798eaf509"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.731750 4936 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.750685 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bc98fe2f-d705-4a0a-83e2-0f5798eaf509" (UID: "bc98fe2f-d705-4a0a-83e2-0f5798eaf509"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.833491 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.833588 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.833617 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.833649 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.833712 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc98fe2f-d705-4a0a-83e2-0f5798eaf509-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.835308 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.835585 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.837662 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.849884 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.853400 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:57 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:57 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:57 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.853469 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.861994 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.870460 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.880199 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:57 crc kubenswrapper[4936]: I0320 16:02:57.883282 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.122804 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.129945 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.167272 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.247034 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd","Type":"ContainerStarted","Data":"e0040e08c64c6aaffc15c6a8aaa5ee155a61efffa5b596fad4ae284448ed7af9"} Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.254522 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc98fe2f-d705-4a0a-83e2-0f5798eaf509","Type":"ContainerDied","Data":"9d58262437b86bc7d7a0016012869f0a465ce2352e063b5c5bcb363bb30b9792"} Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.254610 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d58262437b86bc7d7a0016012869f0a465ce2352e063b5c5bcb363bb30b9792" Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.254751 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.268675 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.268656541 podStartE2EDuration="3.268656541s" podCreationTimestamp="2026-03-20 16:02:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:02:58.265653658 +0000 UTC m=+129.212021463" watchObservedRunningTime="2026-03-20 16:02:58.268656541 +0000 UTC m=+129.215024346" Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.847444 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.852597 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-jgfjb" Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.854864 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:58 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:58 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:58 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:58 crc kubenswrapper[4936]: I0320 16:02:58.854925 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.004302 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.275861 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f78dc0c88838761480edf57781baea8febe1c2b9346bf91553d7d736bc2f7b32"} Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.284868 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"690f14b211a2a5419ab2cc19a53b8f96e55094cced6e40bbb682efe966e0d259"} Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.288837 4936 generic.go:334] "Generic (PLEG): container finished" podID="23ed9f1c-439a-4626-a8e3-5c804cfd6bcd" containerID="e0040e08c64c6aaffc15c6a8aaa5ee155a61efffa5b596fad4ae284448ed7af9" exitCode=0 Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.288892 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd","Type":"ContainerDied","Data":"e0040e08c64c6aaffc15c6a8aaa5ee155a61efffa5b596fad4ae284448ed7af9"} Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.295174 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4b52538bdd51ec02b3b074a9fc7032c1a6ab3da095465b7e6a8753e0c54bc44a"} Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.314862 4936 ???:1] "http: TLS handshake error from 192.168.126.11:35494: no serving certificate available for the kubelet" Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.553655 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-5nbl4" Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.854177 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:02:59 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:02:59 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:02:59 crc kubenswrapper[4936]: healthz check failed Mar 20 16:02:59 crc kubenswrapper[4936]: I0320 16:02:59.854329 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.356344 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"70adf666192f190355591db9602a0654a7f6d10bc2e79dc16e2465c85e105e22"} Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.370897 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8cc9648d04df01dcbb281d8a250697b4b9968325b0816ce23518d8ebee87169d"} Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.388923 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4ccfed756411439d1f6802b3b1ee80c5ee9d0a99d01afcac92dab72c2f90453d"} Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.704232 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.791454 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kubelet-dir\") pod \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\" (UID: \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\") " Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.791626 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kube-api-access\") pod \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\" (UID: \"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd\") " Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.791611 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "23ed9f1c-439a-4626-a8e3-5c804cfd6bcd" (UID: "23ed9f1c-439a-4626-a8e3-5c804cfd6bcd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.792194 4936 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.806966 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "23ed9f1c-439a-4626-a8e3-5c804cfd6bcd" (UID: "23ed9f1c-439a-4626-a8e3-5c804cfd6bcd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.853312 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:03:00 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:03:00 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:03:00 crc kubenswrapper[4936]: healthz check failed Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.853390 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:03:00 crc kubenswrapper[4936]: I0320 16:03:00.893400 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23ed9f1c-439a-4626-a8e3-5c804cfd6bcd-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:01 crc kubenswrapper[4936]: I0320 16:03:01.412811 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 16:03:01 crc kubenswrapper[4936]: I0320 16:03:01.412651 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"23ed9f1c-439a-4626-a8e3-5c804cfd6bcd","Type":"ContainerDied","Data":"a823287493fbfa065447b454e3b60a18d335fbfc6c4dac2f8b2f0cedf4443d54"} Mar 20 16:03:01 crc kubenswrapper[4936]: I0320 16:03:01.414330 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a823287493fbfa065447b454e3b60a18d335fbfc6c4dac2f8b2f0cedf4443d54" Mar 20 16:03:01 crc kubenswrapper[4936]: I0320 16:03:01.414381 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:03:01 crc kubenswrapper[4936]: I0320 16:03:01.856287 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:03:01 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:03:01 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:03:01 crc kubenswrapper[4936]: healthz check failed Mar 20 16:03:01 crc kubenswrapper[4936]: I0320 16:03:01.856364 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:03:02 crc kubenswrapper[4936]: I0320 16:03:02.853594 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:03:02 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:03:02 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:03:02 crc kubenswrapper[4936]: healthz check failed Mar 20 16:03:02 crc kubenswrapper[4936]: I0320 16:03:02.853695 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:03:03 crc kubenswrapper[4936]: I0320 16:03:03.638376 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-dzgl2" Mar 20 16:03:03 crc kubenswrapper[4936]: E0320 16:03:03.844888 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:03 crc kubenswrapper[4936]: E0320 16:03:03.848416 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:03 crc kubenswrapper[4936]: I0320 16:03:03.853040 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:03:03 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:03:03 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:03:03 crc kubenswrapper[4936]: healthz check failed Mar 20 16:03:03 crc kubenswrapper[4936]: I0320 16:03:03.853135 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:03:03 crc kubenswrapper[4936]: E0320 16:03:03.853564 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:03 crc kubenswrapper[4936]: E0320 16:03:03.853613 4936 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerName="kube-multus-additional-cni-plugins" Mar 20 16:03:03 crc kubenswrapper[4936]: I0320 16:03:03.952068 4936 patch_prober.go:28] interesting pod/console-f9d7485db-6q5fl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 20 16:03:03 crc kubenswrapper[4936]: I0320 16:03:03.952123 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6q5fl" podUID="c52c0329-ca68-4d95-8341-ab8191c3a186" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 20 16:03:04 crc kubenswrapper[4936]: I0320 16:03:04.852846 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:03:04 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:03:04 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:03:04 crc kubenswrapper[4936]: healthz check failed Mar 20 16:03:04 crc kubenswrapper[4936]: I0320 16:03:04.852935 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:03:05 crc kubenswrapper[4936]: I0320 16:03:05.856426 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:03:05 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:03:05 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:03:05 crc kubenswrapper[4936]: healthz check failed Mar 20 16:03:05 crc kubenswrapper[4936]: I0320 16:03:05.856962 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:03:06 crc kubenswrapper[4936]: I0320 16:03:06.853381 4936 patch_prober.go:28] interesting pod/router-default-5444994796-9ltqh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 16:03:06 crc kubenswrapper[4936]: [-]has-synced failed: reason withheld Mar 20 16:03:06 crc kubenswrapper[4936]: [+]process-running ok Mar 20 16:03:06 crc kubenswrapper[4936]: healthz check failed Mar 20 16:03:06 crc kubenswrapper[4936]: I0320 16:03:06.853438 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9ltqh" podUID="1766c31e-3eb1-4723-9fa9-4768f826f8ec" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 16:03:07 crc kubenswrapper[4936]: I0320 16:03:07.871690 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:03:07 crc kubenswrapper[4936]: I0320 16:03:07.878806 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-9ltqh" Mar 20 16:03:10 crc kubenswrapper[4936]: I0320 16:03:10.516888 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-559b965cdc-6sctl"] Mar 20 16:03:10 crc kubenswrapper[4936]: I0320 16:03:10.517445 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" podUID="ff75dce3-7056-4983-99d8-f67217bdba2f" containerName="controller-manager" containerID="cri-o://f56f07ae6c26c400f7195662e3f6cd991836ba2513903775ffed98ca15c86937" gracePeriod=30 Mar 20 16:03:10 crc kubenswrapper[4936]: I0320 16:03:10.552491 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h"] Mar 20 16:03:10 crc kubenswrapper[4936]: I0320 16:03:10.552830 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" podUID="67ff8a83-b028-4b8f-9fee-58847d147c08" containerName="route-controller-manager" containerID="cri-o://dd242b81c4774344d9415e73f518ad3898125c58bb421ddbcac6f967fd2bea6e" gracePeriod=30 Mar 20 16:03:10 crc kubenswrapper[4936]: E0320 16:03:10.622860 4936 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff75dce3_7056_4983_99d8_f67217bdba2f.slice/crio-f56f07ae6c26c400f7195662e3f6cd991836ba2513903775ffed98ca15c86937.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67ff8a83_b028_4b8f_9fee_58847d147c08.slice/crio-dd242b81c4774344d9415e73f518ad3898125c58bb421ddbcac6f967fd2bea6e.scope\": RecentStats: unable to find data in memory cache]" Mar 20 16:03:11 crc kubenswrapper[4936]: I0320 16:03:11.507060 4936 generic.go:334] "Generic (PLEG): container finished" podID="ff75dce3-7056-4983-99d8-f67217bdba2f" containerID="f56f07ae6c26c400f7195662e3f6cd991836ba2513903775ffed98ca15c86937" exitCode=0 Mar 20 16:03:11 crc kubenswrapper[4936]: I0320 16:03:11.507200 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" event={"ID":"ff75dce3-7056-4983-99d8-f67217bdba2f","Type":"ContainerDied","Data":"f56f07ae6c26c400f7195662e3f6cd991836ba2513903775ffed98ca15c86937"} Mar 20 16:03:11 crc kubenswrapper[4936]: I0320 16:03:11.509817 4936 generic.go:334] "Generic (PLEG): container finished" podID="67ff8a83-b028-4b8f-9fee-58847d147c08" containerID="dd242b81c4774344d9415e73f518ad3898125c58bb421ddbcac6f967fd2bea6e" exitCode=0 Mar 20 16:03:11 crc kubenswrapper[4936]: I0320 16:03:11.509868 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" event={"ID":"67ff8a83-b028-4b8f-9fee-58847d147c08","Type":"ContainerDied","Data":"dd242b81c4774344d9415e73f518ad3898125c58bb421ddbcac6f967fd2bea6e"} Mar 20 16:03:12 crc kubenswrapper[4936]: I0320 16:03:12.562855 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:03:13 crc kubenswrapper[4936]: I0320 16:03:13.325014 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:03:13 crc kubenswrapper[4936]: I0320 16:03:13.705975 4936 patch_prober.go:28] interesting pod/route-controller-manager-5989f75564-7hm4h container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" start-of-body= Mar 20 16:03:13 crc kubenswrapper[4936]: I0320 16:03:13.706084 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" podUID="67ff8a83-b028-4b8f-9fee-58847d147c08" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" Mar 20 16:03:13 crc kubenswrapper[4936]: I0320 16:03:13.712363 4936 patch_prober.go:28] interesting pod/controller-manager-559b965cdc-6sctl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" start-of-body= Mar 20 16:03:13 crc kubenswrapper[4936]: I0320 16:03:13.712514 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" podUID="ff75dce3-7056-4983-99d8-f67217bdba2f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" Mar 20 16:03:13 crc kubenswrapper[4936]: E0320 16:03:13.838557 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:13 crc kubenswrapper[4936]: E0320 16:03:13.841148 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:13 crc kubenswrapper[4936]: E0320 16:03:13.842788 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:13 crc kubenswrapper[4936]: E0320 16:03:13.842824 4936 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerName="kube-multus-additional-cni-plugins" Mar 20 16:03:13 crc kubenswrapper[4936]: I0320 16:03:13.875669 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 20 16:03:13 crc kubenswrapper[4936]: I0320 16:03:13.955365 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:03:13 crc kubenswrapper[4936]: I0320 16:03:13.959236 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:03:14 crc kubenswrapper[4936]: I0320 16:03:14.006714 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=1.006684975 podStartE2EDuration="1.006684975s" podCreationTimestamp="2026-03-20 16:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:03:14.00650529 +0000 UTC m=+144.952873115" watchObservedRunningTime="2026-03-20 16:03:14.006684975 +0000 UTC m=+144.953052790" Mar 20 16:03:18 crc kubenswrapper[4936]: I0320 16:03:18.875469 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 20 16:03:19 crc kubenswrapper[4936]: I0320 16:03:19.832390 4936 ???:1] "http: TLS handshake error from 192.168.126.11:45684: no serving certificate available for the kubelet" Mar 20 16:03:19 crc kubenswrapper[4936]: I0320 16:03:19.874995 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.874973295 podStartE2EDuration="1.874973295s" podCreationTimestamp="2026-03-20 16:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:03:19.871001534 +0000 UTC m=+150.817369369" watchObservedRunningTime="2026-03-20 16:03:19.874973295 +0000 UTC m=+150.821341110" Mar 20 16:03:20 crc kubenswrapper[4936]: W0320 16:03:20.808902 4936 helpers.go:245] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab39813c_c2d7_432c_aa05_7b33c1e86fac.slice/crio-df16c32748a31661bfdf4facd428473b6c6913c2602b8fa785b82ef4edc30bb3/memory.swap.max": read /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab39813c_c2d7_432c_aa05_7b33c1e86fac.slice/crio-df16c32748a31661bfdf4facd428473b6c6913c2602b8fa785b82ef4edc30bb3/memory.swap.max: no such device Mar 20 16:03:21 crc kubenswrapper[4936]: I0320 16:03:21.597900 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-m5rqk_ab39813c-c2d7-432c-aa05-7b33c1e86fac/kube-multus-additional-cni-plugins/0.log" Mar 20 16:03:21 crc kubenswrapper[4936]: I0320 16:03:21.598010 4936 generic.go:334] "Generic (PLEG): container finished" podID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" exitCode=137 Mar 20 16:03:21 crc kubenswrapper[4936]: I0320 16:03:21.598082 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" event={"ID":"ab39813c-c2d7-432c-aa05-7b33c1e86fac","Type":"ContainerDied","Data":"57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf"} Mar 20 16:03:23 crc kubenswrapper[4936]: E0320 16:03:23.837005 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf is running failed: container process not found" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:23 crc kubenswrapper[4936]: E0320 16:03:23.838241 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf is running failed: container process not found" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:23 crc kubenswrapper[4936]: E0320 16:03:23.838630 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf is running failed: container process not found" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:23 crc kubenswrapper[4936]: E0320 16:03:23.838669 4936 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerName="kube-multus-additional-cni-plugins" Mar 20 16:03:24 crc kubenswrapper[4936]: I0320 16:03:24.704936 4936 patch_prober.go:28] interesting pod/route-controller-manager-5989f75564-7hm4h container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:03:24 crc kubenswrapper[4936]: I0320 16:03:24.705075 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" podUID="67ff8a83-b028-4b8f-9fee-58847d147c08" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 16:03:24 crc kubenswrapper[4936]: I0320 16:03:24.713468 4936 patch_prober.go:28] interesting pod/controller-manager-559b965cdc-6sctl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:03:24 crc kubenswrapper[4936]: I0320 16:03:24.713582 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" podUID="ff75dce3-7056-4983-99d8-f67217bdba2f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 16:03:24 crc kubenswrapper[4936]: I0320 16:03:24.847974 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wspgg" Mar 20 16:03:26 crc kubenswrapper[4936]: E0320 16:03:26.255769 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 20 16:03:26 crc kubenswrapper[4936]: E0320 16:03:26.256001 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5rwpr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pqjxj_openshift-marketplace(cd71509d-889b-498c-b5ec-55e6902b08f7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 16:03:26 crc kubenswrapper[4936]: E0320 16:03:26.257227 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pqjxj" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.408048 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 16:03:28 crc kubenswrapper[4936]: E0320 16:03:28.408326 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ed9f1c-439a-4626-a8e3-5c804cfd6bcd" containerName="pruner" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.408344 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ed9f1c-439a-4626-a8e3-5c804cfd6bcd" containerName="pruner" Mar 20 16:03:28 crc kubenswrapper[4936]: E0320 16:03:28.408366 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc98fe2f-d705-4a0a-83e2-0f5798eaf509" containerName="pruner" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.408373 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc98fe2f-d705-4a0a-83e2-0f5798eaf509" containerName="pruner" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.408486 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ed9f1c-439a-4626-a8e3-5c804cfd6bcd" containerName="pruner" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.408500 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc98fe2f-d705-4a0a-83e2-0f5798eaf509" containerName="pruner" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.408908 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.416414 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.421922 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.428712 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.592340 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.592421 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.693764 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.693906 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.694054 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.723980 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:28 crc kubenswrapper[4936]: I0320 16:03:28.744735 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:32 crc kubenswrapper[4936]: E0320 16:03:32.898558 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-pqjxj" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" Mar 20 16:03:32 crc kubenswrapper[4936]: E0320 16:03:32.972861 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 20 16:03:32 crc kubenswrapper[4936]: E0320 16:03:32.973107 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qlwzp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6xpds_openshift-marketplace(9b3ba1a3-fe09-48fe-a584-3cbd20c2225a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 16:03:32 crc kubenswrapper[4936]: E0320 16:03:32.974523 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-6xpds" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" Mar 20 16:03:33 crc kubenswrapper[4936]: E0320 16:03:33.838008 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf is running failed: container process not found" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:33 crc kubenswrapper[4936]: E0320 16:03:33.838911 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf is running failed: container process not found" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:33 crc kubenswrapper[4936]: E0320 16:03:33.839499 4936 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf is running failed: container process not found" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 20 16:03:33 crc kubenswrapper[4936]: E0320 16:03:33.839566 4936 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerName="kube-multus-additional-cni-plugins" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.704746 4936 patch_prober.go:28] interesting pod/route-controller-manager-5989f75564-7hm4h container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": context deadline exceeded" start-of-body= Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.705078 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" podUID="67ff8a83-b028-4b8f-9fee-58847d147c08" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": context deadline exceeded" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.711048 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6xpds" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.713140 4936 patch_prober.go:28] interesting pod/controller-manager-559b965cdc-6sctl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.713205 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" podUID="ff75dce3-7056-4983-99d8-f67217bdba2f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.801085 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.801281 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gjhn7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xp5nm_openshift-marketplace(a649556e-fd89-4edb-89f7-d1db999c616b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.803816 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xp5nm" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.804112 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.804520 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sp7h4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9fw5r_openshift-marketplace(014c12c3-e743-45cd-b916-875bf90f7e00): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.805693 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9fw5r" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.832163 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.832535 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.832562 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.832775 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dj5v4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-njqd9_openshift-marketplace(959c1ced-7501-4754-8e17-c3a420edd1b8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.832918 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nhdlt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kg66h_openshift-marketplace(b7e7487c-e0c4-47a9-9875-b1c41e14eecc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.834640 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kg66h" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.834531 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-njqd9" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.858465 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.875245 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6fd8f8579c-76b9b"] Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.877973 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ff8a83-b028-4b8f-9fee-58847d147c08" containerName="route-controller-manager" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.877988 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ff8a83-b028-4b8f-9fee-58847d147c08" containerName="route-controller-manager" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.878001 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff75dce3-7056-4983-99d8-f67217bdba2f" containerName="controller-manager" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.878008 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff75dce3-7056-4983-99d8-f67217bdba2f" containerName="controller-manager" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.878164 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff75dce3-7056-4983-99d8-f67217bdba2f" containerName="controller-manager" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.878181 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ff8a83-b028-4b8f-9fee-58847d147c08" containerName="route-controller-manager" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.879977 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.881724 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-m5rqk_ab39813c-c2d7-432c-aa05-7b33c1e86fac/kube-multus-additional-cni-plugins/0.log" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.881769 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.893149 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6fd8f8579c-76b9b"] Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.899270 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.899536 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vd9t9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5cpcn_openshift-marketplace(7964eb91-819c-4853-a598-9b93346f496f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 16:03:34 crc kubenswrapper[4936]: E0320 16:03:34.901910 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5cpcn" podUID="7964eb91-819c-4853-a598-9b93346f496f" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.974144 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-proxy-ca-bundles\") pod \"ff75dce3-7056-4983-99d8-f67217bdba2f\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.974217 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c6hh\" (UniqueName: \"kubernetes.io/projected/ab39813c-c2d7-432c-aa05-7b33c1e86fac-kube-api-access-5c6hh\") pod \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.974256 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-client-ca\") pod \"ff75dce3-7056-4983-99d8-f67217bdba2f\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.975231 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab39813c-c2d7-432c-aa05-7b33c1e86fac-tuning-conf-dir\") pod \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.974923 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ff75dce3-7056-4983-99d8-f67217bdba2f" (UID: "ff75dce3-7056-4983-99d8-f67217bdba2f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.975156 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-client-ca" (OuterVolumeSpecName: "client-ca") pod "ff75dce3-7056-4983-99d8-f67217bdba2f" (UID: "ff75dce3-7056-4983-99d8-f67217bdba2f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.975325 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8phk\" (UniqueName: \"kubernetes.io/projected/ff75dce3-7056-4983-99d8-f67217bdba2f-kube-api-access-k8phk\") pod \"ff75dce3-7056-4983-99d8-f67217bdba2f\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.975433 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab39813c-c2d7-432c-aa05-7b33c1e86fac-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "ab39813c-c2d7-432c-aa05-7b33c1e86fac" (UID: "ab39813c-c2d7-432c-aa05-7b33c1e86fac"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.975805 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff75dce3-7056-4983-99d8-f67217bdba2f-serving-cert\") pod \"ff75dce3-7056-4983-99d8-f67217bdba2f\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.975845 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ff8a83-b028-4b8f-9fee-58847d147c08-serving-cert\") pod \"67ff8a83-b028-4b8f-9fee-58847d147c08\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.975902 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-client-ca\") pod \"67ff8a83-b028-4b8f-9fee-58847d147c08\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.975939 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-config\") pod \"67ff8a83-b028-4b8f-9fee-58847d147c08\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.975981 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-config\") pod \"ff75dce3-7056-4983-99d8-f67217bdba2f\" (UID: \"ff75dce3-7056-4983-99d8-f67217bdba2f\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.976011 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab39813c-c2d7-432c-aa05-7b33c1e86fac-cni-sysctl-allowlist\") pod \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.976061 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ab39813c-c2d7-432c-aa05-7b33c1e86fac-ready\") pod \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\" (UID: \"ab39813c-c2d7-432c-aa05-7b33c1e86fac\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.976086 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmh2z\" (UniqueName: \"kubernetes.io/projected/67ff8a83-b028-4b8f-9fee-58847d147c08-kube-api-access-zmh2z\") pod \"67ff8a83-b028-4b8f-9fee-58847d147c08\" (UID: \"67ff8a83-b028-4b8f-9fee-58847d147c08\") " Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.976944 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.976964 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.976977 4936 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab39813c-c2d7-432c-aa05-7b33c1e86fac-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.977868 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-client-ca" (OuterVolumeSpecName: "client-ca") pod "67ff8a83-b028-4b8f-9fee-58847d147c08" (UID: "67ff8a83-b028-4b8f-9fee-58847d147c08"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.978321 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab39813c-c2d7-432c-aa05-7b33c1e86fac-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "ab39813c-c2d7-432c-aa05-7b33c1e86fac" (UID: "ab39813c-c2d7-432c-aa05-7b33c1e86fac"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.978444 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab39813c-c2d7-432c-aa05-7b33c1e86fac-ready" (OuterVolumeSpecName: "ready") pod "ab39813c-c2d7-432c-aa05-7b33c1e86fac" (UID: "ab39813c-c2d7-432c-aa05-7b33c1e86fac"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.978487 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-config" (OuterVolumeSpecName: "config") pod "ff75dce3-7056-4983-99d8-f67217bdba2f" (UID: "ff75dce3-7056-4983-99d8-f67217bdba2f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.979370 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-config" (OuterVolumeSpecName: "config") pod "67ff8a83-b028-4b8f-9fee-58847d147c08" (UID: "67ff8a83-b028-4b8f-9fee-58847d147c08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.983720 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff75dce3-7056-4983-99d8-f67217bdba2f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ff75dce3-7056-4983-99d8-f67217bdba2f" (UID: "ff75dce3-7056-4983-99d8-f67217bdba2f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.983740 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67ff8a83-b028-4b8f-9fee-58847d147c08-kube-api-access-zmh2z" (OuterVolumeSpecName: "kube-api-access-zmh2z") pod "67ff8a83-b028-4b8f-9fee-58847d147c08" (UID: "67ff8a83-b028-4b8f-9fee-58847d147c08"). InnerVolumeSpecName "kube-api-access-zmh2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.983833 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff75dce3-7056-4983-99d8-f67217bdba2f-kube-api-access-k8phk" (OuterVolumeSpecName: "kube-api-access-k8phk") pod "ff75dce3-7056-4983-99d8-f67217bdba2f" (UID: "ff75dce3-7056-4983-99d8-f67217bdba2f"). InnerVolumeSpecName "kube-api-access-k8phk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.984730 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ff8a83-b028-4b8f-9fee-58847d147c08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "67ff8a83-b028-4b8f-9fee-58847d147c08" (UID: "67ff8a83-b028-4b8f-9fee-58847d147c08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:03:34 crc kubenswrapper[4936]: I0320 16:03:34.988128 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab39813c-c2d7-432c-aa05-7b33c1e86fac-kube-api-access-5c6hh" (OuterVolumeSpecName: "kube-api-access-5c6hh") pod "ab39813c-c2d7-432c-aa05-7b33c1e86fac" (UID: "ab39813c-c2d7-432c-aa05-7b33c1e86fac"). InnerVolumeSpecName "kube-api-access-5c6hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078204 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srh9h\" (UniqueName: \"kubernetes.io/projected/50a66a8a-8305-4949-a1c1-09848668b7bc-kube-api-access-srh9h\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078317 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-client-ca\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078348 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-config\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078372 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50a66a8a-8305-4949-a1c1-09848668b7bc-serving-cert\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078414 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-proxy-ca-bundles\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078467 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c6hh\" (UniqueName: \"kubernetes.io/projected/ab39813c-c2d7-432c-aa05-7b33c1e86fac-kube-api-access-5c6hh\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078480 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8phk\" (UniqueName: \"kubernetes.io/projected/ff75dce3-7056-4983-99d8-f67217bdba2f-kube-api-access-k8phk\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078491 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff75dce3-7056-4983-99d8-f67217bdba2f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078501 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ff8a83-b028-4b8f-9fee-58847d147c08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078512 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078521 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ff8a83-b028-4b8f-9fee-58847d147c08-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078530 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff75dce3-7056-4983-99d8-f67217bdba2f-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078542 4936 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab39813c-c2d7-432c-aa05-7b33c1e86fac-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078567 4936 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ab39813c-c2d7-432c-aa05-7b33c1e86fac-ready\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.078577 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmh2z\" (UniqueName: \"kubernetes.io/projected/67ff8a83-b028-4b8f-9fee-58847d147c08-kube-api-access-zmh2z\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.180324 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-config\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.180428 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50a66a8a-8305-4949-a1c1-09848668b7bc-serving-cert\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.180512 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-proxy-ca-bundles\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.180588 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srh9h\" (UniqueName: \"kubernetes.io/projected/50a66a8a-8305-4949-a1c1-09848668b7bc-kube-api-access-srh9h\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.180716 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-client-ca\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.181889 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-proxy-ca-bundles\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.182676 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-config\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.182727 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-client-ca\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.186529 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50a66a8a-8305-4949-a1c1-09848668b7bc-serving-cert\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.197302 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srh9h\" (UniqueName: \"kubernetes.io/projected/50a66a8a-8305-4949-a1c1-09848668b7bc-kube-api-access-srh9h\") pod \"controller-manager-6fd8f8579c-76b9b\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.212252 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 16:03:35 crc kubenswrapper[4936]: W0320 16:03:35.219671 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2aefdd27_9e99_4131_9b3e_70f63b1d98cb.slice/crio-f59f2270619f546a7735f1ab7541fef22319403acd802f7b2a688732a3179046 WatchSource:0}: Error finding container f59f2270619f546a7735f1ab7541fef22319403acd802f7b2a688732a3179046: Status 404 returned error can't find the container with id f59f2270619f546a7735f1ab7541fef22319403acd802f7b2a688732a3179046 Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.494084 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.691654 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6fd8f8579c-76b9b"] Mar 20 16:03:35 crc kubenswrapper[4936]: W0320 16:03:35.708633 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50a66a8a_8305_4949_a1c1_09848668b7bc.slice/crio-04e027c0525217436341eb2b534241123fbd87694f90ad1065f42afabf55f554 WatchSource:0}: Error finding container 04e027c0525217436341eb2b534241123fbd87694f90ad1065f42afabf55f554: Status 404 returned error can't find the container with id 04e027c0525217436341eb2b534241123fbd87694f90ad1065f42afabf55f554 Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.711319 4936 generic.go:334] "Generic (PLEG): container finished" podID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerID="6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de" exitCode=0 Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.711422 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km59s" event={"ID":"2aae37a3-f585-4b9b-94e3-5560e06da84c","Type":"ContainerDied","Data":"6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de"} Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.713450 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2aefdd27-9e99-4131-9b3e-70f63b1d98cb","Type":"ContainerStarted","Data":"63210c65279bcc0d02757fa812d40b8756577a98a72f2d0c32cb836eb7ac8b54"} Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.713533 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2aefdd27-9e99-4131-9b3e-70f63b1d98cb","Type":"ContainerStarted","Data":"f59f2270619f546a7735f1ab7541fef22319403acd802f7b2a688732a3179046"} Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.717917 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" event={"ID":"67ff8a83-b028-4b8f-9fee-58847d147c08","Type":"ContainerDied","Data":"c490d4a14c8c59730e495ce35969b205b7349bdd344ba7b380605935baf7b12c"} Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.717993 4936 scope.go:117] "RemoveContainer" containerID="dd242b81c4774344d9415e73f518ad3898125c58bb421ddbcac6f967fd2bea6e" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.718171 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.727403 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-m5rqk_ab39813c-c2d7-432c-aa05-7b33c1e86fac/kube-multus-additional-cni-plugins/0.log" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.727576 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" event={"ID":"ab39813c-c2d7-432c-aa05-7b33c1e86fac","Type":"ContainerDied","Data":"df16c32748a31661bfdf4facd428473b6c6913c2602b8fa785b82ef4edc30bb3"} Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.727723 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-m5rqk" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.740380 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" event={"ID":"ff75dce3-7056-4983-99d8-f67217bdba2f","Type":"ContainerDied","Data":"d51847f92beee533240ab9a08d9020b49d4d6b470ba05e406e47b2aa4fcc5438"} Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.740279 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-559b965cdc-6sctl" Mar 20 16:03:35 crc kubenswrapper[4936]: E0320 16:03:35.759616 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xp5nm" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" Mar 20 16:03:35 crc kubenswrapper[4936]: E0320 16:03:35.760802 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kg66h" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.774846 4936 scope.go:117] "RemoveContainer" containerID="57ce3a522c2a254a9c7cd8819d27c02e8dbaccb86e3461729b5db28ad3054dbf" Mar 20 16:03:35 crc kubenswrapper[4936]: E0320 16:03:35.775879 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-njqd9" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" Mar 20 16:03:35 crc kubenswrapper[4936]: E0320 16:03:35.778775 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5cpcn" podUID="7964eb91-819c-4853-a598-9b93346f496f" Mar 20 16:03:35 crc kubenswrapper[4936]: E0320 16:03:35.780978 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9fw5r" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.786731 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=7.786706662 podStartE2EDuration="7.786706662s" podCreationTimestamp="2026-03-20 16:03:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:03:35.776201704 +0000 UTC m=+166.722569519" watchObservedRunningTime="2026-03-20 16:03:35.786706662 +0000 UTC m=+166.733074477" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.818616 4936 scope.go:117] "RemoveContainer" containerID="f56f07ae6c26c400f7195662e3f6cd991836ba2513903775ffed98ca15c86937" Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.886586 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h"] Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.889956 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5989f75564-7hm4h"] Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.898392 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-559b965cdc-6sctl"] Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.904456 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-559b965cdc-6sctl"] Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.910025 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-m5rqk"] Mar 20 16:03:35 crc kubenswrapper[4936]: I0320 16:03:35.914188 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-m5rqk"] Mar 20 16:03:36 crc kubenswrapper[4936]: I0320 16:03:36.748534 4936 generic.go:334] "Generic (PLEG): container finished" podID="2aefdd27-9e99-4131-9b3e-70f63b1d98cb" containerID="63210c65279bcc0d02757fa812d40b8756577a98a72f2d0c32cb836eb7ac8b54" exitCode=0 Mar 20 16:03:36 crc kubenswrapper[4936]: I0320 16:03:36.748599 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2aefdd27-9e99-4131-9b3e-70f63b1d98cb","Type":"ContainerDied","Data":"63210c65279bcc0d02757fa812d40b8756577a98a72f2d0c32cb836eb7ac8b54"} Mar 20 16:03:36 crc kubenswrapper[4936]: I0320 16:03:36.751055 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" event={"ID":"50a66a8a-8305-4949-a1c1-09848668b7bc","Type":"ContainerStarted","Data":"61e14f601a774116626fc5212715d6e1121d5302abc3c811f5e9f1bc3e0df9a7"} Mar 20 16:03:36 crc kubenswrapper[4936]: I0320 16:03:36.751122 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" event={"ID":"50a66a8a-8305-4949-a1c1-09848668b7bc","Type":"ContainerStarted","Data":"04e027c0525217436341eb2b534241123fbd87694f90ad1065f42afabf55f554"} Mar 20 16:03:36 crc kubenswrapper[4936]: I0320 16:03:36.751157 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:36 crc kubenswrapper[4936]: I0320 16:03:36.768937 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:03:36 crc kubenswrapper[4936]: I0320 16:03:36.770191 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km59s" event={"ID":"2aae37a3-f585-4b9b-94e3-5560e06da84c","Type":"ContainerStarted","Data":"ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb"} Mar 20 16:03:36 crc kubenswrapper[4936]: I0320 16:03:36.800031 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-km59s" podStartSLOduration=4.136079662 podStartE2EDuration="47.800002748s" podCreationTimestamp="2026-03-20 16:02:49 +0000 UTC" firstStartedPulling="2026-03-20 16:02:52.650896816 +0000 UTC m=+123.597264631" lastFinishedPulling="2026-03-20 16:03:36.314819892 +0000 UTC m=+167.261187717" observedRunningTime="2026-03-20 16:03:36.798667744 +0000 UTC m=+167.745035559" watchObservedRunningTime="2026-03-20 16:03:36.800002748 +0000 UTC m=+167.746370583" Mar 20 16:03:36 crc kubenswrapper[4936]: I0320 16:03:36.818050 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" podStartSLOduration=6.818019307 podStartE2EDuration="6.818019307s" podCreationTimestamp="2026-03-20 16:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:03:36.815704048 +0000 UTC m=+167.762071903" watchObservedRunningTime="2026-03-20 16:03:36.818019307 +0000 UTC m=+167.764387152" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.385272 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc"] Mar 20 16:03:37 crc kubenswrapper[4936]: E0320 16:03:37.386323 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerName="kube-multus-additional-cni-plugins" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.386399 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerName="kube-multus-additional-cni-plugins" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.386637 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" containerName="kube-multus-additional-cni-plugins" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.387207 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.390464 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.390738 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.393138 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.393325 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.399139 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.399685 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.408659 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc"] Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.517250 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-client-ca\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.517423 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v59nl\" (UniqueName: \"kubernetes.io/projected/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-kube-api-access-v59nl\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.517590 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-serving-cert\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.517671 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-config\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.619147 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-serving-cert\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.619228 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-config\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.619268 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-client-ca\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.619299 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v59nl\" (UniqueName: \"kubernetes.io/projected/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-kube-api-access-v59nl\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.620583 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-client-ca\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.620770 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-config\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.627358 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-serving-cert\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.637280 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v59nl\" (UniqueName: \"kubernetes.io/projected/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-kube-api-access-v59nl\") pod \"route-controller-manager-7db6d8559-9f9xc\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.709157 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.862571 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67ff8a83-b028-4b8f-9fee-58847d147c08" path="/var/lib/kubelet/pods/67ff8a83-b028-4b8f-9fee-58847d147c08/volumes" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.864149 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab39813c-c2d7-432c-aa05-7b33c1e86fac" path="/var/lib/kubelet/pods/ab39813c-c2d7-432c-aa05-7b33c1e86fac/volumes" Mar 20 16:03:37 crc kubenswrapper[4936]: I0320 16:03:37.865241 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff75dce3-7056-4983-99d8-f67217bdba2f" path="/var/lib/kubelet/pods/ff75dce3-7056-4983-99d8-f67217bdba2f/volumes" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.066603 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.135175 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.220629 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc"] Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.226820 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kubelet-dir\") pod \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\" (UID: \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\") " Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.226970 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kube-api-access\") pod \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\" (UID: \"2aefdd27-9e99-4131-9b3e-70f63b1d98cb\") " Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.227108 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2aefdd27-9e99-4131-9b3e-70f63b1d98cb" (UID: "2aefdd27-9e99-4131-9b3e-70f63b1d98cb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.227397 4936 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.231020 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2aefdd27-9e99-4131-9b3e-70f63b1d98cb" (UID: "2aefdd27-9e99-4131-9b3e-70f63b1d98cb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.329385 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2aefdd27-9e99-4131-9b3e-70f63b1d98cb-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.785027 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" event={"ID":"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0","Type":"ContainerStarted","Data":"4ddcfb8452f379d62a83214ebcc474dba675c143f507744d6b6f982b47cebf46"} Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.785090 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" event={"ID":"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0","Type":"ContainerStarted","Data":"b48df8c3e2f65753efc0deb2e578f7c19b2e061556dbf7c78722ac10f58d7f27"} Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.785414 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.788374 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2aefdd27-9e99-4131-9b3e-70f63b1d98cb","Type":"ContainerDied","Data":"f59f2270619f546a7735f1ab7541fef22319403acd802f7b2a688732a3179046"} Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.788429 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f59f2270619f546a7735f1ab7541fef22319403acd802f7b2a688732a3179046" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.788438 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.806164 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" podStartSLOduration=8.806134586 podStartE2EDuration="8.806134586s" podCreationTimestamp="2026-03-20 16:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:03:38.804678049 +0000 UTC m=+169.751045864" watchObservedRunningTime="2026-03-20 16:03:38.806134586 +0000 UTC m=+169.752502401" Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.963024 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jhqvm"] Mar 20 16:03:38 crc kubenswrapper[4936]: I0320 16:03:38.985792 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:03:40 crc kubenswrapper[4936]: I0320 16:03:40.253119 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-km59s" Mar 20 16:03:40 crc kubenswrapper[4936]: I0320 16:03:40.253670 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-km59s" Mar 20 16:03:40 crc kubenswrapper[4936]: I0320 16:03:40.432165 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-km59s" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.220363 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 16:03:43 crc kubenswrapper[4936]: E0320 16:03:43.220955 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aefdd27-9e99-4131-9b3e-70f63b1d98cb" containerName="pruner" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.220973 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aefdd27-9e99-4131-9b3e-70f63b1d98cb" containerName="pruner" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.221080 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aefdd27-9e99-4131-9b3e-70f63b1d98cb" containerName="pruner" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.221530 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.227890 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.228560 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.232220 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.400499 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-var-lock\") pod \"installer-9-crc\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.400642 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-kubelet-dir\") pod \"installer-9-crc\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.400682 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87e9a750-951a-40e5-b714-e128787afa82-kube-api-access\") pod \"installer-9-crc\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.501585 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-var-lock\") pod \"installer-9-crc\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.501691 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-kubelet-dir\") pod \"installer-9-crc\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.501706 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-var-lock\") pod \"installer-9-crc\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.501734 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87e9a750-951a-40e5-b714-e128787afa82-kube-api-access\") pod \"installer-9-crc\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.501864 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-kubelet-dir\") pod \"installer-9-crc\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.525651 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87e9a750-951a-40e5-b714-e128787afa82-kube-api-access\") pod \"installer-9-crc\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.553107 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:03:43 crc kubenswrapper[4936]: I0320 16:03:43.981929 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 16:03:43 crc kubenswrapper[4936]: W0320 16:03:43.996405 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod87e9a750_951a_40e5_b714_e128787afa82.slice/crio-1e35e0ff39b17839a6df55182a480f18a282095d8c3bbba77f2c9c2ddd2f513e WatchSource:0}: Error finding container 1e35e0ff39b17839a6df55182a480f18a282095d8c3bbba77f2c9c2ddd2f513e: Status 404 returned error can't find the container with id 1e35e0ff39b17839a6df55182a480f18a282095d8c3bbba77f2c9c2ddd2f513e Mar 20 16:03:44 crc kubenswrapper[4936]: I0320 16:03:44.827831 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"87e9a750-951a-40e5-b714-e128787afa82","Type":"ContainerStarted","Data":"fc6d08a42d139f322b64022a1e8003e18664364d819f5a80b9f6a623906e2a31"} Mar 20 16:03:44 crc kubenswrapper[4936]: I0320 16:03:44.827894 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"87e9a750-951a-40e5-b714-e128787afa82","Type":"ContainerStarted","Data":"1e35e0ff39b17839a6df55182a480f18a282095d8c3bbba77f2c9c2ddd2f513e"} Mar 20 16:03:44 crc kubenswrapper[4936]: I0320 16:03:44.862218 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.8621967929999999 podStartE2EDuration="1.862196793s" podCreationTimestamp="2026-03-20 16:03:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:03:44.860673034 +0000 UTC m=+175.807040849" watchObservedRunningTime="2026-03-20 16:03:44.862196793 +0000 UTC m=+175.808564618" Mar 20 16:03:47 crc kubenswrapper[4936]: I0320 16:03:47.849720 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njqd9" event={"ID":"959c1ced-7501-4754-8e17-c3a420edd1b8","Type":"ContainerStarted","Data":"3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56"} Mar 20 16:03:47 crc kubenswrapper[4936]: I0320 16:03:47.885335 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg66h" event={"ID":"b7e7487c-e0c4-47a9-9875-b1c41e14eecc","Type":"ContainerStarted","Data":"3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e"} Mar 20 16:03:47 crc kubenswrapper[4936]: I0320 16:03:47.885406 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqjxj" event={"ID":"cd71509d-889b-498c-b5ec-55e6902b08f7","Type":"ContainerStarted","Data":"7fed475bcbc3466cf1c0e6068980dd17ea57662d46e77210349fa67dd4b6252b"} Mar 20 16:03:48 crc kubenswrapper[4936]: I0320 16:03:48.884560 4936 generic.go:334] "Generic (PLEG): container finished" podID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerID="3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56" exitCode=0 Mar 20 16:03:48 crc kubenswrapper[4936]: I0320 16:03:48.885277 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njqd9" event={"ID":"959c1ced-7501-4754-8e17-c3a420edd1b8","Type":"ContainerDied","Data":"3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56"} Mar 20 16:03:48 crc kubenswrapper[4936]: I0320 16:03:48.892041 4936 generic.go:334] "Generic (PLEG): container finished" podID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerID="3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e" exitCode=0 Mar 20 16:03:48 crc kubenswrapper[4936]: I0320 16:03:48.892117 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg66h" event={"ID":"b7e7487c-e0c4-47a9-9875-b1c41e14eecc","Type":"ContainerDied","Data":"3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e"} Mar 20 16:03:48 crc kubenswrapper[4936]: I0320 16:03:48.898068 4936 generic.go:334] "Generic (PLEG): container finished" podID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerID="7fed475bcbc3466cf1c0e6068980dd17ea57662d46e77210349fa67dd4b6252b" exitCode=0 Mar 20 16:03:48 crc kubenswrapper[4936]: I0320 16:03:48.898131 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqjxj" event={"ID":"cd71509d-889b-498c-b5ec-55e6902b08f7","Type":"ContainerDied","Data":"7fed475bcbc3466cf1c0e6068980dd17ea57662d46e77210349fa67dd4b6252b"} Mar 20 16:03:48 crc kubenswrapper[4936]: I0320 16:03:48.901584 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fw5r" event={"ID":"014c12c3-e743-45cd-b916-875bf90f7e00","Type":"ContainerStarted","Data":"f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d"} Mar 20 16:03:49 crc kubenswrapper[4936]: I0320 16:03:49.909310 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg66h" event={"ID":"b7e7487c-e0c4-47a9-9875-b1c41e14eecc","Type":"ContainerStarted","Data":"8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1"} Mar 20 16:03:49 crc kubenswrapper[4936]: I0320 16:03:49.911194 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqjxj" event={"ID":"cd71509d-889b-498c-b5ec-55e6902b08f7","Type":"ContainerStarted","Data":"de143d5aa189f9de1f698c73ad88bd726a4dcb5f2a0e898b786b966e52411744"} Mar 20 16:03:49 crc kubenswrapper[4936]: I0320 16:03:49.914521 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xpds" event={"ID":"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a","Type":"ContainerStarted","Data":"f816f8d846737fe42a177f1629753162c76ffc4cee60790d9d38fd7230426c8f"} Mar 20 16:03:49 crc kubenswrapper[4936]: I0320 16:03:49.924735 4936 generic.go:334] "Generic (PLEG): container finished" podID="014c12c3-e743-45cd-b916-875bf90f7e00" containerID="f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d" exitCode=0 Mar 20 16:03:49 crc kubenswrapper[4936]: I0320 16:03:49.924820 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fw5r" event={"ID":"014c12c3-e743-45cd-b916-875bf90f7e00","Type":"ContainerDied","Data":"f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d"} Mar 20 16:03:49 crc kubenswrapper[4936]: I0320 16:03:49.934017 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5cpcn" event={"ID":"7964eb91-819c-4853-a598-9b93346f496f","Type":"ContainerStarted","Data":"7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67"} Mar 20 16:03:49 crc kubenswrapper[4936]: I0320 16:03:49.938436 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pqjxj" podStartSLOduration=3.532861796 podStartE2EDuration="57.938422936s" podCreationTimestamp="2026-03-20 16:02:52 +0000 UTC" firstStartedPulling="2026-03-20 16:02:55.002600011 +0000 UTC m=+125.948967826" lastFinishedPulling="2026-03-20 16:03:49.408161141 +0000 UTC m=+180.354528966" observedRunningTime="2026-03-20 16:03:49.933606563 +0000 UTC m=+180.879974378" watchObservedRunningTime="2026-03-20 16:03:49.938422936 +0000 UTC m=+180.884790751" Mar 20 16:03:49 crc kubenswrapper[4936]: I0320 16:03:49.948857 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njqd9" event={"ID":"959c1ced-7501-4754-8e17-c3a420edd1b8","Type":"ContainerStarted","Data":"e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6"} Mar 20 16:03:50 crc kubenswrapper[4936]: I0320 16:03:50.009646 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-njqd9" podStartSLOduration=4.682174889 podStartE2EDuration="59.009621851s" podCreationTimestamp="2026-03-20 16:02:51 +0000 UTC" firstStartedPulling="2026-03-20 16:02:55.108005675 +0000 UTC m=+126.054373490" lastFinishedPulling="2026-03-20 16:03:49.435452597 +0000 UTC m=+180.381820452" observedRunningTime="2026-03-20 16:03:50.004383817 +0000 UTC m=+180.950751622" watchObservedRunningTime="2026-03-20 16:03:50.009621851 +0000 UTC m=+180.955989666" Mar 20 16:03:50 crc kubenswrapper[4936]: I0320 16:03:50.321595 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-km59s" Mar 20 16:03:50 crc kubenswrapper[4936]: I0320 16:03:50.956176 4936 generic.go:334] "Generic (PLEG): container finished" podID="7964eb91-819c-4853-a598-9b93346f496f" containerID="7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67" exitCode=0 Mar 20 16:03:50 crc kubenswrapper[4936]: I0320 16:03:50.956265 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5cpcn" event={"ID":"7964eb91-819c-4853-a598-9b93346f496f","Type":"ContainerDied","Data":"7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67"} Mar 20 16:03:50 crc kubenswrapper[4936]: I0320 16:03:50.957781 4936 generic.go:334] "Generic (PLEG): container finished" podID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerID="f816f8d846737fe42a177f1629753162c76ffc4cee60790d9d38fd7230426c8f" exitCode=0 Mar 20 16:03:50 crc kubenswrapper[4936]: I0320 16:03:50.957815 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xpds" event={"ID":"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a","Type":"ContainerDied","Data":"f816f8d846737fe42a177f1629753162c76ffc4cee60790d9d38fd7230426c8f"} Mar 20 16:03:51 crc kubenswrapper[4936]: I0320 16:03:51.026505 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kg66h" podStartSLOduration=4.50486137 podStartE2EDuration="59.026478317s" podCreationTimestamp="2026-03-20 16:02:52 +0000 UTC" firstStartedPulling="2026-03-20 16:02:54.992358611 +0000 UTC m=+125.938726426" lastFinishedPulling="2026-03-20 16:03:49.513975558 +0000 UTC m=+180.460343373" observedRunningTime="2026-03-20 16:03:51.022175657 +0000 UTC m=+181.968543472" watchObservedRunningTime="2026-03-20 16:03:51.026478317 +0000 UTC m=+181.972846132" Mar 20 16:03:53 crc kubenswrapper[4936]: I0320 16:03:53.219599 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:03:53 crc kubenswrapper[4936]: I0320 16:03:53.220616 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:03:53 crc kubenswrapper[4936]: I0320 16:03:53.266401 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:03:53 crc kubenswrapper[4936]: I0320 16:03:53.266472 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:03:53 crc kubenswrapper[4936]: I0320 16:03:53.308344 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:03:53 crc kubenswrapper[4936]: I0320 16:03:53.330430 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:03:53 crc kubenswrapper[4936]: I0320 16:03:53.335081 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:03:53 crc kubenswrapper[4936]: I0320 16:03:53.381328 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:03:53 crc kubenswrapper[4936]: I0320 16:03:53.977223 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fw5r" event={"ID":"014c12c3-e743-45cd-b916-875bf90f7e00","Type":"ContainerStarted","Data":"d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe"} Mar 20 16:03:54 crc kubenswrapper[4936]: I0320 16:03:54.258816 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kg66h" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerName="registry-server" probeResult="failure" output=< Mar 20 16:03:54 crc kubenswrapper[4936]: timeout: failed to connect service ":50051" within 1s Mar 20 16:03:54 crc kubenswrapper[4936]: > Mar 20 16:03:55 crc kubenswrapper[4936]: I0320 16:03:55.016007 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9fw5r" podStartSLOduration=4.670775733 podStartE2EDuration="1m5.015980334s" podCreationTimestamp="2026-03-20 16:02:50 +0000 UTC" firstStartedPulling="2026-03-20 16:02:52.686717916 +0000 UTC m=+123.633085741" lastFinishedPulling="2026-03-20 16:03:53.031922527 +0000 UTC m=+183.978290342" observedRunningTime="2026-03-20 16:03:55.013271855 +0000 UTC m=+185.959639670" watchObservedRunningTime="2026-03-20 16:03:55.015980334 +0000 UTC m=+185.962348149" Mar 20 16:03:55 crc kubenswrapper[4936]: I0320 16:03:55.028460 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:03:56 crc kubenswrapper[4936]: I0320 16:03:56.292980 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqjxj"] Mar 20 16:03:57 crc kubenswrapper[4936]: I0320 16:03:57.007946 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xpds" event={"ID":"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a","Type":"ContainerStarted","Data":"e48a8d760c02e4b6fd6431641f69a93807454684578810940ceb81bba0d42c10"} Mar 20 16:03:57 crc kubenswrapper[4936]: I0320 16:03:57.010358 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pqjxj" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerName="registry-server" containerID="cri-o://de143d5aa189f9de1f698c73ad88bd726a4dcb5f2a0e898b786b966e52411744" gracePeriod=2 Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.022612 4936 generic.go:334] "Generic (PLEG): container finished" podID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerID="de143d5aa189f9de1f698c73ad88bd726a4dcb5f2a0e898b786b966e52411744" exitCode=0 Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.022691 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqjxj" event={"ID":"cd71509d-889b-498c-b5ec-55e6902b08f7","Type":"ContainerDied","Data":"de143d5aa189f9de1f698c73ad88bd726a4dcb5f2a0e898b786b966e52411744"} Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.744024 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.764269 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6xpds" podStartSLOduration=5.557803633 podStartE2EDuration="1m5.764248453s" podCreationTimestamp="2026-03-20 16:02:53 +0000 UTC" firstStartedPulling="2026-03-20 16:02:54.945607372 +0000 UTC m=+125.891975187" lastFinishedPulling="2026-03-20 16:03:55.152052192 +0000 UTC m=+186.098420007" observedRunningTime="2026-03-20 16:03:57.032972249 +0000 UTC m=+187.979340094" watchObservedRunningTime="2026-03-20 16:03:58.764248453 +0000 UTC m=+189.710616268" Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.859472 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-utilities\") pod \"cd71509d-889b-498c-b5ec-55e6902b08f7\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.859578 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-catalog-content\") pod \"cd71509d-889b-498c-b5ec-55e6902b08f7\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.859667 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rwpr\" (UniqueName: \"kubernetes.io/projected/cd71509d-889b-498c-b5ec-55e6902b08f7-kube-api-access-5rwpr\") pod \"cd71509d-889b-498c-b5ec-55e6902b08f7\" (UID: \"cd71509d-889b-498c-b5ec-55e6902b08f7\") " Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.860724 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-utilities" (OuterVolumeSpecName: "utilities") pod "cd71509d-889b-498c-b5ec-55e6902b08f7" (UID: "cd71509d-889b-498c-b5ec-55e6902b08f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.866439 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd71509d-889b-498c-b5ec-55e6902b08f7-kube-api-access-5rwpr" (OuterVolumeSpecName: "kube-api-access-5rwpr") pod "cd71509d-889b-498c-b5ec-55e6902b08f7" (UID: "cd71509d-889b-498c-b5ec-55e6902b08f7"). InnerVolumeSpecName "kube-api-access-5rwpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.898210 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd71509d-889b-498c-b5ec-55e6902b08f7" (UID: "cd71509d-889b-498c-b5ec-55e6902b08f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.961661 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.961690 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd71509d-889b-498c-b5ec-55e6902b08f7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:58 crc kubenswrapper[4936]: I0320 16:03:58.961717 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rwpr\" (UniqueName: \"kubernetes.io/projected/cd71509d-889b-498c-b5ec-55e6902b08f7-kube-api-access-5rwpr\") on node \"crc\" DevicePath \"\"" Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.035463 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqjxj" event={"ID":"cd71509d-889b-498c-b5ec-55e6902b08f7","Type":"ContainerDied","Data":"f7293e3ed6f67cd098b73da8f73f22f6eb1b2c3843a0d8516a747b2d94b31f81"} Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.035520 4936 scope.go:117] "RemoveContainer" containerID="de143d5aa189f9de1f698c73ad88bd726a4dcb5f2a0e898b786b966e52411744" Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.035710 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqjxj" Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.038809 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5cpcn" event={"ID":"7964eb91-819c-4853-a598-9b93346f496f","Type":"ContainerStarted","Data":"110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b"} Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.073800 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5cpcn" podStartSLOduration=5.181466218 podStartE2EDuration="1m10.073785592s" podCreationTimestamp="2026-03-20 16:02:49 +0000 UTC" firstStartedPulling="2026-03-20 16:02:52.67406749 +0000 UTC m=+123.620435305" lastFinishedPulling="2026-03-20 16:03:57.566386854 +0000 UTC m=+188.512754679" observedRunningTime="2026-03-20 16:03:59.069131423 +0000 UTC m=+190.015499238" watchObservedRunningTime="2026-03-20 16:03:59.073785592 +0000 UTC m=+190.020153407" Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.088274 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqjxj"] Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.094065 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqjxj"] Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.417085 4936 scope.go:117] "RemoveContainer" containerID="7fed475bcbc3466cf1c0e6068980dd17ea57662d46e77210349fa67dd4b6252b" Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.450044 4936 scope.go:117] "RemoveContainer" containerID="e22926ace22dfafb0ec423692d3f189059d67ddb67183f81e7c57d62e858ac98" Mar 20 16:03:59 crc kubenswrapper[4936]: I0320 16:03:59.865697 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" path="/var/lib/kubelet/pods/cd71509d-889b-498c-b5ec-55e6902b08f7/volumes" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.049345 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xp5nm" event={"ID":"a649556e-fd89-4edb-89f7-d1db999c616b","Type":"ContainerStarted","Data":"1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0"} Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.060247 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.060654 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.147499 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567044-4hcwr"] Mar 20 16:04:00 crc kubenswrapper[4936]: E0320 16:04:00.147834 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerName="extract-content" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.147864 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerName="extract-content" Mar 20 16:04:00 crc kubenswrapper[4936]: E0320 16:04:00.147900 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerName="extract-utilities" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.147912 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerName="extract-utilities" Mar 20 16:04:00 crc kubenswrapper[4936]: E0320 16:04:00.147937 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerName="registry-server" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.147948 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerName="registry-server" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.148080 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd71509d-889b-498c-b5ec-55e6902b08f7" containerName="registry-server" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.148636 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567044-4hcwr" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.150681 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.153068 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.156389 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567044-4hcwr"] Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.156448 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.183103 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzccx\" (UniqueName: \"kubernetes.io/projected/531e6b04-d999-4aca-ad9d-981c1db726d9-kube-api-access-pzccx\") pod \"auto-csr-approver-29567044-4hcwr\" (UID: \"531e6b04-d999-4aca-ad9d-981c1db726d9\") " pod="openshift-infra/auto-csr-approver-29567044-4hcwr" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.284557 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzccx\" (UniqueName: \"kubernetes.io/projected/531e6b04-d999-4aca-ad9d-981c1db726d9-kube-api-access-pzccx\") pod \"auto-csr-approver-29567044-4hcwr\" (UID: \"531e6b04-d999-4aca-ad9d-981c1db726d9\") " pod="openshift-infra/auto-csr-approver-29567044-4hcwr" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.303112 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzccx\" (UniqueName: \"kubernetes.io/projected/531e6b04-d999-4aca-ad9d-981c1db726d9-kube-api-access-pzccx\") pod \"auto-csr-approver-29567044-4hcwr\" (UID: \"531e6b04-d999-4aca-ad9d-981c1db726d9\") " pod="openshift-infra/auto-csr-approver-29567044-4hcwr" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.466395 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567044-4hcwr" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.770585 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.770923 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.812741 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.826082 4936 ???:1] "http: TLS handshake error from 192.168.126.11:56084: no serving certificate available for the kubelet" Mar 20 16:04:00 crc kubenswrapper[4936]: I0320 16:04:00.944373 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567044-4hcwr"] Mar 20 16:04:00 crc kubenswrapper[4936]: W0320 16:04:00.950098 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod531e6b04_d999_4aca_ad9d_981c1db726d9.slice/crio-75bc61bf571f2840b3c0c467f03a59e2dec94cecfb577c1f110d4ca811a201a8 WatchSource:0}: Error finding container 75bc61bf571f2840b3c0c467f03a59e2dec94cecfb577c1f110d4ca811a201a8: Status 404 returned error can't find the container with id 75bc61bf571f2840b3c0c467f03a59e2dec94cecfb577c1f110d4ca811a201a8 Mar 20 16:04:01 crc kubenswrapper[4936]: I0320 16:04:01.056508 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567044-4hcwr" event={"ID":"531e6b04-d999-4aca-ad9d-981c1db726d9","Type":"ContainerStarted","Data":"75bc61bf571f2840b3c0c467f03a59e2dec94cecfb577c1f110d4ca811a201a8"} Mar 20 16:04:01 crc kubenswrapper[4936]: I0320 16:04:01.058232 4936 generic.go:334] "Generic (PLEG): container finished" podID="a649556e-fd89-4edb-89f7-d1db999c616b" containerID="1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0" exitCode=0 Mar 20 16:04:01 crc kubenswrapper[4936]: I0320 16:04:01.058299 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xp5nm" event={"ID":"a649556e-fd89-4edb-89f7-d1db999c616b","Type":"ContainerDied","Data":"1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0"} Mar 20 16:04:01 crc kubenswrapper[4936]: I0320 16:04:01.109825 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:04:01 crc kubenswrapper[4936]: I0320 16:04:01.115308 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-5cpcn" podUID="7964eb91-819c-4853-a598-9b93346f496f" containerName="registry-server" probeResult="failure" output=< Mar 20 16:04:01 crc kubenswrapper[4936]: timeout: failed to connect service ":50051" within 1s Mar 20 16:04:01 crc kubenswrapper[4936]: > Mar 20 16:04:02 crc kubenswrapper[4936]: I0320 16:04:02.068790 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xp5nm" event={"ID":"a649556e-fd89-4edb-89f7-d1db999c616b","Type":"ContainerStarted","Data":"ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a"} Mar 20 16:04:02 crc kubenswrapper[4936]: I0320 16:04:02.096526 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9fw5r"] Mar 20 16:04:02 crc kubenswrapper[4936]: I0320 16:04:02.100085 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xp5nm" podStartSLOduration=3.097781546 podStartE2EDuration="1m12.10006978s" podCreationTimestamp="2026-03-20 16:02:50 +0000 UTC" firstStartedPulling="2026-03-20 16:02:52.640222024 +0000 UTC m=+123.586589839" lastFinishedPulling="2026-03-20 16:04:01.642510218 +0000 UTC m=+192.588878073" observedRunningTime="2026-03-20 16:04:02.091267026 +0000 UTC m=+193.037634861" watchObservedRunningTime="2026-03-20 16:04:02.10006978 +0000 UTC m=+193.046437595" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.075193 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9fw5r" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" containerName="registry-server" containerID="cri-o://d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe" gracePeriod=2 Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.272583 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.308769 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.347873 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.613109 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.613276 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.668101 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.703748 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.836348 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp7h4\" (UniqueName: \"kubernetes.io/projected/014c12c3-e743-45cd-b916-875bf90f7e00-kube-api-access-sp7h4\") pod \"014c12c3-e743-45cd-b916-875bf90f7e00\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.836502 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-utilities\") pod \"014c12c3-e743-45cd-b916-875bf90f7e00\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.836607 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-catalog-content\") pod \"014c12c3-e743-45cd-b916-875bf90f7e00\" (UID: \"014c12c3-e743-45cd-b916-875bf90f7e00\") " Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.837840 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-utilities" (OuterVolumeSpecName: "utilities") pod "014c12c3-e743-45cd-b916-875bf90f7e00" (UID: "014c12c3-e743-45cd-b916-875bf90f7e00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.846860 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/014c12c3-e743-45cd-b916-875bf90f7e00-kube-api-access-sp7h4" (OuterVolumeSpecName: "kube-api-access-sp7h4") pod "014c12c3-e743-45cd-b916-875bf90f7e00" (UID: "014c12c3-e743-45cd-b916-875bf90f7e00"). InnerVolumeSpecName "kube-api-access-sp7h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.891448 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "014c12c3-e743-45cd-b916-875bf90f7e00" (UID: "014c12c3-e743-45cd-b916-875bf90f7e00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.938895 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.938956 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp7h4\" (UniqueName: \"kubernetes.io/projected/014c12c3-e743-45cd-b916-875bf90f7e00-kube-api-access-sp7h4\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:03 crc kubenswrapper[4936]: I0320 16:04:03.938973 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/014c12c3-e743-45cd-b916-875bf90f7e00-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.011433 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" podUID="9333cf9d-b80d-4242-a940-c92ed3b593d1" containerName="oauth-openshift" containerID="cri-o://c7e27bbd84991bdee07cd35e5752be306d3cd838ff6bdddba0e68de90443db8b" gracePeriod=15 Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.087843 4936 generic.go:334] "Generic (PLEG): container finished" podID="014c12c3-e743-45cd-b916-875bf90f7e00" containerID="d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe" exitCode=0 Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.087965 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fw5r" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.088089 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fw5r" event={"ID":"014c12c3-e743-45cd-b916-875bf90f7e00","Type":"ContainerDied","Data":"d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe"} Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.088154 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fw5r" event={"ID":"014c12c3-e743-45cd-b916-875bf90f7e00","Type":"ContainerDied","Data":"4cf01fd44fb32e1558e4fee63b79fbd849f8c89d019d20112213cf946203ea74"} Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.088194 4936 scope.go:117] "RemoveContainer" containerID="d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.129003 4936 scope.go:117] "RemoveContainer" containerID="f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.146811 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9fw5r"] Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.155047 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9fw5r"] Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.158489 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.169844 4936 scope.go:117] "RemoveContainer" containerID="109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.199802 4936 scope.go:117] "RemoveContainer" containerID="d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe" Mar 20 16:04:04 crc kubenswrapper[4936]: E0320 16:04:04.202370 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe\": container with ID starting with d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe not found: ID does not exist" containerID="d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.202436 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe"} err="failed to get container status \"d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe\": rpc error: code = NotFound desc = could not find container \"d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe\": container with ID starting with d25d2ca3acaf730aafb59238f8c35de83cb6d55ed2f82ffd737c69b1aa2152fe not found: ID does not exist" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.202475 4936 scope.go:117] "RemoveContainer" containerID="f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d" Mar 20 16:04:04 crc kubenswrapper[4936]: E0320 16:04:04.203120 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d\": container with ID starting with f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d not found: ID does not exist" containerID="f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.203190 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d"} err="failed to get container status \"f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d\": rpc error: code = NotFound desc = could not find container \"f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d\": container with ID starting with f91ee929d970dd035da3c1f152c0a3a12aa34945da513ece98dffbc24909274d not found: ID does not exist" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.203236 4936 scope.go:117] "RemoveContainer" containerID="109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88" Mar 20 16:04:04 crc kubenswrapper[4936]: E0320 16:04:04.203839 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88\": container with ID starting with 109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88 not found: ID does not exist" containerID="109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88" Mar 20 16:04:04 crc kubenswrapper[4936]: I0320 16:04:04.203870 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88"} err="failed to get container status \"109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88\": rpc error: code = NotFound desc = could not find container \"109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88\": container with ID starting with 109ae8fb61b4eb3ed246550d03f7faf6aca0ac62b0a797fa3ec4243809758f88 not found: ID does not exist" Mar 20 16:04:05 crc kubenswrapper[4936]: I0320 16:04:05.098093 4936 generic.go:334] "Generic (PLEG): container finished" podID="9333cf9d-b80d-4242-a940-c92ed3b593d1" containerID="c7e27bbd84991bdee07cd35e5752be306d3cd838ff6bdddba0e68de90443db8b" exitCode=0 Mar 20 16:04:05 crc kubenswrapper[4936]: I0320 16:04:05.098189 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" event={"ID":"9333cf9d-b80d-4242-a940-c92ed3b593d1","Type":"ContainerDied","Data":"c7e27bbd84991bdee07cd35e5752be306d3cd838ff6bdddba0e68de90443db8b"} Mar 20 16:04:05 crc kubenswrapper[4936]: I0320 16:04:05.863471 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" path="/var/lib/kubelet/pods/014c12c3-e743-45cd-b916-875bf90f7e00/volumes" Mar 20 16:04:06 crc kubenswrapper[4936]: I0320 16:04:06.491138 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6xpds"] Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.107289 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.112596 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" event={"ID":"9333cf9d-b80d-4242-a940-c92ed3b593d1","Type":"ContainerDied","Data":"edb98d9b986602424f3f6097e9cb7ebe0c6da2e355696f64ca8bc1b3100b78c8"} Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.112650 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jhqvm" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.112706 4936 scope.go:117] "RemoveContainer" containerID="c7e27bbd84991bdee07cd35e5752be306d3cd838ff6bdddba0e68de90443db8b" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.112838 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6xpds" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerName="registry-server" containerID="cri-o://e48a8d760c02e4b6fd6431641f69a93807454684578810940ceb81bba0d42c10" gracePeriod=2 Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.194647 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-serving-cert\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.194787 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-error\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196056 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-cliconfig\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196136 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-trusted-ca-bundle\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196229 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-session\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196279 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-router-certs\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196323 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-login\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196376 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-policies\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196445 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-provider-selection\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196538 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-ocp-branding-template\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196604 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-service-ca\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196638 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-dir\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196683 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg5hx\" (UniqueName: \"kubernetes.io/projected/9333cf9d-b80d-4242-a940-c92ed3b593d1-kube-api-access-rg5hx\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.196713 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-idp-0-file-data\") pod \"9333cf9d-b80d-4242-a940-c92ed3b593d1\" (UID: \"9333cf9d-b80d-4242-a940-c92ed3b593d1\") " Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.198184 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.200667 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.201272 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.205156 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.208455 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.212723 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.213072 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.213262 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.220341 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.221047 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.222695 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.223299 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.223483 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9333cf9d-b80d-4242-a940-c92ed3b593d1-kube-api-access-rg5hx" (OuterVolumeSpecName: "kube-api-access-rg5hx") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "kube-api-access-rg5hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.223572 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "9333cf9d-b80d-4242-a940-c92ed3b593d1" (UID: "9333cf9d-b80d-4242-a940-c92ed3b593d1"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.298886 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.298937 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.298951 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.298964 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.298980 4936 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.298995 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.299006 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.299017 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.299028 4936 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9333cf9d-b80d-4242-a940-c92ed3b593d1-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.299038 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg5hx\" (UniqueName: \"kubernetes.io/projected/9333cf9d-b80d-4242-a940-c92ed3b593d1-kube-api-access-rg5hx\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.299047 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.299057 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.299066 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.299075 4936 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9333cf9d-b80d-4242-a940-c92ed3b593d1-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.452373 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jhqvm"] Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.455295 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jhqvm"] Mar 20 16:04:07 crc kubenswrapper[4936]: I0320 16:04:07.864885 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9333cf9d-b80d-4242-a940-c92ed3b593d1" path="/var/lib/kubelet/pods/9333cf9d-b80d-4242-a940-c92ed3b593d1/volumes" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.122095 4936 generic.go:334] "Generic (PLEG): container finished" podID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerID="e48a8d760c02e4b6fd6431641f69a93807454684578810940ceb81bba0d42c10" exitCode=0 Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.122156 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xpds" event={"ID":"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a","Type":"ContainerDied","Data":"e48a8d760c02e4b6fd6431641f69a93807454684578810940ceb81bba0d42c10"} Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.415475 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-64597b9cc5-x6zm8"] Mar 20 16:04:08 crc kubenswrapper[4936]: E0320 16:04:08.415852 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" containerName="extract-content" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.415871 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" containerName="extract-content" Mar 20 16:04:08 crc kubenswrapper[4936]: E0320 16:04:08.415883 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" containerName="registry-server" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.415890 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" containerName="registry-server" Mar 20 16:04:08 crc kubenswrapper[4936]: E0320 16:04:08.415918 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" containerName="extract-utilities" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.415928 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" containerName="extract-utilities" Mar 20 16:04:08 crc kubenswrapper[4936]: E0320 16:04:08.415943 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9333cf9d-b80d-4242-a940-c92ed3b593d1" containerName="oauth-openshift" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.415950 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="9333cf9d-b80d-4242-a940-c92ed3b593d1" containerName="oauth-openshift" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.416072 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="9333cf9d-b80d-4242-a940-c92ed3b593d1" containerName="oauth-openshift" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.416088 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="014c12c3-e743-45cd-b916-875bf90f7e00" containerName="registry-server" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.416744 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.418411 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.420724 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.422518 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.427314 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.427405 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.427885 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.428215 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.428397 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.428419 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.429014 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.429149 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.434468 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.446054 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.447411 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.460418 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64597b9cc5-x6zm8"] Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.460679 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.617736 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-audit-policies\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.617838 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/105ea192-b386-4c6a-836b-6fe5d0572a57-audit-dir\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.617866 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-template-login\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.617895 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.617921 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.617936 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-session\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.617960 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-router-certs\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.617982 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-template-error\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.618772 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5jjg\" (UniqueName: \"kubernetes.io/projected/105ea192-b386-4c6a-836b-6fe5d0572a57-kube-api-access-n5jjg\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.618901 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.618952 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.619001 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.619035 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-service-ca\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.619068 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.719901 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.719971 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-audit-policies\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720011 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/105ea192-b386-4c6a-836b-6fe5d0572a57-audit-dir\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720029 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-template-login\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720050 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720074 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720095 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-session\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720115 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-router-certs\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720132 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-template-error\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720156 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5jjg\" (UniqueName: \"kubernetes.io/projected/105ea192-b386-4c6a-836b-6fe5d0572a57-kube-api-access-n5jjg\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720182 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720186 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/105ea192-b386-4c6a-836b-6fe5d0572a57-audit-dir\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720211 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720316 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720443 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-service-ca\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.720931 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.721097 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-audit-policies\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.721243 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.721386 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-service-ca\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.730580 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-session\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.730655 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.730740 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.731155 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.752039 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-template-error\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.752183 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-router-certs\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.752740 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-user-template-login\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.752936 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/105ea192-b386-4c6a-836b-6fe5d0572a57-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.758273 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5jjg\" (UniqueName: \"kubernetes.io/projected/105ea192-b386-4c6a-836b-6fe5d0572a57-kube-api-access-n5jjg\") pod \"oauth-openshift-64597b9cc5-x6zm8\" (UID: \"105ea192-b386-4c6a-836b-6fe5d0572a57\") " pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.795197 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.821080 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlwzp\" (UniqueName: \"kubernetes.io/projected/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-kube-api-access-qlwzp\") pod \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.821173 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-catalog-content\") pod \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.821216 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-utilities\") pod \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\" (UID: \"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a\") " Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.822044 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-utilities" (OuterVolumeSpecName: "utilities") pod "9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" (UID: "9b3ba1a3-fe09-48fe-a584-3cbd20c2225a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.829855 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-kube-api-access-qlwzp" (OuterVolumeSpecName: "kube-api-access-qlwzp") pod "9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" (UID: "9b3ba1a3-fe09-48fe-a584-3cbd20c2225a"). InnerVolumeSpecName "kube-api-access-qlwzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.931104 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:08 crc kubenswrapper[4936]: I0320 16:04:08.931192 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlwzp\" (UniqueName: \"kubernetes.io/projected/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-kube-api-access-qlwzp\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.013266 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" (UID: "9b3ba1a3-fe09-48fe-a584-3cbd20c2225a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.034163 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.047908 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.139996 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xpds" event={"ID":"9b3ba1a3-fe09-48fe-a584-3cbd20c2225a","Type":"ContainerDied","Data":"5901809f687d4eb42800894c2d8bd5fe947ad4f155d60efbfce75e4ba4b72c70"} Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.140067 4936 scope.go:117] "RemoveContainer" containerID="e48a8d760c02e4b6fd6431641f69a93807454684578810940ceb81bba0d42c10" Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.140074 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xpds" Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.173754 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6xpds"] Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.175185 4936 scope.go:117] "RemoveContainer" containerID="f816f8d846737fe42a177f1629753162c76ffc4cee60790d9d38fd7230426c8f" Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.183170 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6xpds"] Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.199595 4936 scope.go:117] "RemoveContainer" containerID="61c8748b50301280d495ee41d29e10d6a82c7d901f6e4fcb8d45d45679fd0ffc" Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.481881 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64597b9cc5-x6zm8"] Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.564861 4936 csr.go:261] certificate signing request csr-m79qf is approved, waiting to be issued Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.572619 4936 csr.go:257] certificate signing request csr-m79qf is issued Mar 20 16:04:09 crc kubenswrapper[4936]: I0320 16:04:09.861218 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" path="/var/lib/kubelet/pods/9b3ba1a3-fe09-48fe-a584-3cbd20c2225a/volumes" Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.117463 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.157069 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" event={"ID":"105ea192-b386-4c6a-836b-6fe5d0572a57","Type":"ContainerStarted","Data":"b0bc058f7346a021e73ced7889ae1de3cad0ba10539c2406a43c12ca70c971ae"} Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.157117 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" event={"ID":"105ea192-b386-4c6a-836b-6fe5d0572a57","Type":"ContainerStarted","Data":"74fc532f46ba48fcdc5d1656a3dccda8221b79c3a18cc88e0d73b4f76764e7aa"} Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.157604 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.161272 4936 generic.go:334] "Generic (PLEG): container finished" podID="531e6b04-d999-4aca-ad9d-981c1db726d9" containerID="0e948211251ab8f49c63f1a5565d2cc9c5d22b946421def1f9e82877b318eaf3" exitCode=0 Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.161319 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567044-4hcwr" event={"ID":"531e6b04-d999-4aca-ad9d-981c1db726d9","Type":"ContainerDied","Data":"0e948211251ab8f49c63f1a5565d2cc9c5d22b946421def1f9e82877b318eaf3"} Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.169468 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.196354 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" podStartSLOduration=32.196328096 podStartE2EDuration="32.196328096s" podCreationTimestamp="2026-03-20 16:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:04:10.19345742 +0000 UTC m=+201.139825255" watchObservedRunningTime="2026-03-20 16:04:10.196328096 +0000 UTC m=+201.142695911" Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.400214 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-64597b9cc5-x6zm8" Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.532307 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6fd8f8579c-76b9b"] Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.532638 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" podUID="50a66a8a-8305-4949-a1c1-09848668b7bc" containerName="controller-manager" containerID="cri-o://61e14f601a774116626fc5212715d6e1121d5302abc3c811f5e9f1bc3e0df9a7" gracePeriod=30 Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.574320 4936 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-05 05:02:11.515150283 +0000 UTC Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.574384 4936 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6972h58m0.940770485s for next certificate rotation Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.637667 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc"] Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.638055 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" podUID="9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" containerName="route-controller-manager" containerID="cri-o://4ddcfb8452f379d62a83214ebcc474dba675c143f507744d6b6f982b47cebf46" gracePeriod=30 Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.704002 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.704086 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:04:10 crc kubenswrapper[4936]: I0320 16:04:10.756940 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.171258 4936 generic.go:334] "Generic (PLEG): container finished" podID="50a66a8a-8305-4949-a1c1-09848668b7bc" containerID="61e14f601a774116626fc5212715d6e1121d5302abc3c811f5e9f1bc3e0df9a7" exitCode=0 Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.171364 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" event={"ID":"50a66a8a-8305-4949-a1c1-09848668b7bc","Type":"ContainerDied","Data":"61e14f601a774116626fc5212715d6e1121d5302abc3c811f5e9f1bc3e0df9a7"} Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.175428 4936 generic.go:334] "Generic (PLEG): container finished" podID="9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" containerID="4ddcfb8452f379d62a83214ebcc474dba675c143f507744d6b6f982b47cebf46" exitCode=0 Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.175606 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" event={"ID":"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0","Type":"ContainerDied","Data":"4ddcfb8452f379d62a83214ebcc474dba675c143f507744d6b6f982b47cebf46"} Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.237635 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.458981 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.482037 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-config\") pod \"50a66a8a-8305-4949-a1c1-09848668b7bc\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.482114 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50a66a8a-8305-4949-a1c1-09848668b7bc-serving-cert\") pod \"50a66a8a-8305-4949-a1c1-09848668b7bc\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.482257 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srh9h\" (UniqueName: \"kubernetes.io/projected/50a66a8a-8305-4949-a1c1-09848668b7bc-kube-api-access-srh9h\") pod \"50a66a8a-8305-4949-a1c1-09848668b7bc\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.482476 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-proxy-ca-bundles\") pod \"50a66a8a-8305-4949-a1c1-09848668b7bc\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.482499 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-client-ca\") pod \"50a66a8a-8305-4949-a1c1-09848668b7bc\" (UID: \"50a66a8a-8305-4949-a1c1-09848668b7bc\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.483369 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-client-ca" (OuterVolumeSpecName: "client-ca") pod "50a66a8a-8305-4949-a1c1-09848668b7bc" (UID: "50a66a8a-8305-4949-a1c1-09848668b7bc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.483436 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-config" (OuterVolumeSpecName: "config") pod "50a66a8a-8305-4949-a1c1-09848668b7bc" (UID: "50a66a8a-8305-4949-a1c1-09848668b7bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.483818 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "50a66a8a-8305-4949-a1c1-09848668b7bc" (UID: "50a66a8a-8305-4949-a1c1-09848668b7bc"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.495270 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a66a8a-8305-4949-a1c1-09848668b7bc-kube-api-access-srh9h" (OuterVolumeSpecName: "kube-api-access-srh9h") pod "50a66a8a-8305-4949-a1c1-09848668b7bc" (UID: "50a66a8a-8305-4949-a1c1-09848668b7bc"). InnerVolumeSpecName "kube-api-access-srh9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.495933 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50a66a8a-8305-4949-a1c1-09848668b7bc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "50a66a8a-8305-4949-a1c1-09848668b7bc" (UID: "50a66a8a-8305-4949-a1c1-09848668b7bc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.499307 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567044-4hcwr" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.557965 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.575068 4936 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-27 19:09:14.622186455 +0000 UTC Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.575115 4936 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6771h5m3.047074038s for next certificate rotation Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585179 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v59nl\" (UniqueName: \"kubernetes.io/projected/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-kube-api-access-v59nl\") pod \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585251 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-client-ca\") pod \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585284 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-serving-cert\") pod \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585353 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-config\") pod \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\" (UID: \"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585435 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzccx\" (UniqueName: \"kubernetes.io/projected/531e6b04-d999-4aca-ad9d-981c1db726d9-kube-api-access-pzccx\") pod \"531e6b04-d999-4aca-ad9d-981c1db726d9\" (UID: \"531e6b04-d999-4aca-ad9d-981c1db726d9\") " Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585826 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50a66a8a-8305-4949-a1c1-09848668b7bc-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585856 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srh9h\" (UniqueName: \"kubernetes.io/projected/50a66a8a-8305-4949-a1c1-09848668b7bc-kube-api-access-srh9h\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585870 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585881 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.585892 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50a66a8a-8305-4949-a1c1-09848668b7bc-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.586564 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-client-ca" (OuterVolumeSpecName: "client-ca") pod "9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" (UID: "9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.586707 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-config" (OuterVolumeSpecName: "config") pod "9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" (UID: "9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.588963 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" (UID: "9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.589004 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-kube-api-access-v59nl" (OuterVolumeSpecName: "kube-api-access-v59nl") pod "9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" (UID: "9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0"). InnerVolumeSpecName "kube-api-access-v59nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.589267 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/531e6b04-d999-4aca-ad9d-981c1db726d9-kube-api-access-pzccx" (OuterVolumeSpecName: "kube-api-access-pzccx") pod "531e6b04-d999-4aca-ad9d-981c1db726d9" (UID: "531e6b04-d999-4aca-ad9d-981c1db726d9"). InnerVolumeSpecName "kube-api-access-pzccx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.687532 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzccx\" (UniqueName: \"kubernetes.io/projected/531e6b04-d999-4aca-ad9d-981c1db726d9-kube-api-access-pzccx\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.687588 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v59nl\" (UniqueName: \"kubernetes.io/projected/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-kube-api-access-v59nl\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.687605 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.687624 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:11 crc kubenswrapper[4936]: I0320 16:04:11.688483 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.184969 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567044-4hcwr" event={"ID":"531e6b04-d999-4aca-ad9d-981c1db726d9","Type":"ContainerDied","Data":"75bc61bf571f2840b3c0c467f03a59e2dec94cecfb577c1f110d4ca811a201a8"} Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.185048 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75bc61bf571f2840b3c0c467f03a59e2dec94cecfb577c1f110d4ca811a201a8" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.184998 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567044-4hcwr" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.187643 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" event={"ID":"50a66a8a-8305-4949-a1c1-09848668b7bc","Type":"ContainerDied","Data":"04e027c0525217436341eb2b534241123fbd87694f90ad1065f42afabf55f554"} Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.187698 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fd8f8579c-76b9b" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.187720 4936 scope.go:117] "RemoveContainer" containerID="61e14f601a774116626fc5212715d6e1121d5302abc3c811f5e9f1bc3e0df9a7" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.191906 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" event={"ID":"9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0","Type":"ContainerDied","Data":"b48df8c3e2f65753efc0deb2e578f7c19b2e061556dbf7c78722ac10f58d7f27"} Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.192075 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.215935 4936 scope.go:117] "RemoveContainer" containerID="4ddcfb8452f379d62a83214ebcc474dba675c143f507744d6b6f982b47cebf46" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.229858 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc"] Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.232712 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db6d8559-9f9xc"] Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.249730 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6fd8f8579c-76b9b"] Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.254326 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6fd8f8579c-76b9b"] Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.427770 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7977b8c78-h7dxg"] Mar 20 16:04:12 crc kubenswrapper[4936]: E0320 16:04:12.428130 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a66a8a-8305-4949-a1c1-09848668b7bc" containerName="controller-manager" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428152 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a66a8a-8305-4949-a1c1-09848668b7bc" containerName="controller-manager" Mar 20 16:04:12 crc kubenswrapper[4936]: E0320 16:04:12.428178 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" containerName="route-controller-manager" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428188 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" containerName="route-controller-manager" Mar 20 16:04:12 crc kubenswrapper[4936]: E0320 16:04:12.428205 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerName="extract-utilities" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428214 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerName="extract-utilities" Mar 20 16:04:12 crc kubenswrapper[4936]: E0320 16:04:12.428227 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerName="extract-content" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428234 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerName="extract-content" Mar 20 16:04:12 crc kubenswrapper[4936]: E0320 16:04:12.428244 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerName="registry-server" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428251 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerName="registry-server" Mar 20 16:04:12 crc kubenswrapper[4936]: E0320 16:04:12.428273 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531e6b04-d999-4aca-ad9d-981c1db726d9" containerName="oc" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428281 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="531e6b04-d999-4aca-ad9d-981c1db726d9" containerName="oc" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428402 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="531e6b04-d999-4aca-ad9d-981c1db726d9" containerName="oc" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428414 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b3ba1a3-fe09-48fe-a584-3cbd20c2225a" containerName="registry-server" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428429 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a66a8a-8305-4949-a1c1-09848668b7bc" containerName="controller-manager" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.428451 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" containerName="route-controller-manager" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.429007 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.432841 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.433351 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.433676 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.433858 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.434020 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.434124 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt"] Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.434514 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.435833 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.438945 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.439052 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.439688 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.440005 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.440069 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7977b8c78-h7dxg"] Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.440329 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.441297 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.445878 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.451819 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt"] Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.504122 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vhrb\" (UniqueName: \"kubernetes.io/projected/084dfad4-4a1b-41b8-87fa-693db016d891-kube-api-access-5vhrb\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.504185 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lspvq\" (UniqueName: \"kubernetes.io/projected/dacda504-4ad0-4da5-9f22-39cc7a32982b-kube-api-access-lspvq\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.504226 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-proxy-ca-bundles\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.504277 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/084dfad4-4a1b-41b8-87fa-693db016d891-serving-cert\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.504326 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-config\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.504358 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacda504-4ad0-4da5-9f22-39cc7a32982b-serving-cert\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.504605 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-client-ca\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.504757 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-config\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.504785 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-client-ca\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.606422 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-client-ca\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.606531 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-config\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.606579 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-client-ca\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.606632 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vhrb\" (UniqueName: \"kubernetes.io/projected/084dfad4-4a1b-41b8-87fa-693db016d891-kube-api-access-5vhrb\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.606658 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lspvq\" (UniqueName: \"kubernetes.io/projected/dacda504-4ad0-4da5-9f22-39cc7a32982b-kube-api-access-lspvq\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.606683 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-proxy-ca-bundles\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.606708 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/084dfad4-4a1b-41b8-87fa-693db016d891-serving-cert\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.606746 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-config\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.606771 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacda504-4ad0-4da5-9f22-39cc7a32982b-serving-cert\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.608141 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-client-ca\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.608186 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-client-ca\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.608309 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-config\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.609622 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-proxy-ca-bundles\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.611082 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-config\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.613964 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/084dfad4-4a1b-41b8-87fa-693db016d891-serving-cert\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.615144 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacda504-4ad0-4da5-9f22-39cc7a32982b-serving-cert\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.631370 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lspvq\" (UniqueName: \"kubernetes.io/projected/dacda504-4ad0-4da5-9f22-39cc7a32982b-kube-api-access-lspvq\") pod \"controller-manager-7977b8c78-h7dxg\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.631595 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vhrb\" (UniqueName: \"kubernetes.io/projected/084dfad4-4a1b-41b8-87fa-693db016d891-kube-api-access-5vhrb\") pod \"route-controller-manager-79bf99c7bf-zqszt\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.691572 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xp5nm"] Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.761762 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:12 crc kubenswrapper[4936]: I0320 16:04:12.771579 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.069770 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7977b8c78-h7dxg"] Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.198647 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" event={"ID":"dacda504-4ad0-4da5-9f22-39cc7a32982b","Type":"ContainerStarted","Data":"2b24b331aaf2b35fa684308804d426b732f5059bd7ccb86211e7210bba283d21"} Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.202657 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xp5nm" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" containerName="registry-server" containerID="cri-o://ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a" gracePeriod=2 Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.233807 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt"] Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.607964 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.721894 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjhn7\" (UniqueName: \"kubernetes.io/projected/a649556e-fd89-4edb-89f7-d1db999c616b-kube-api-access-gjhn7\") pod \"a649556e-fd89-4edb-89f7-d1db999c616b\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.723302 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-utilities\") pod \"a649556e-fd89-4edb-89f7-d1db999c616b\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.723349 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-catalog-content\") pod \"a649556e-fd89-4edb-89f7-d1db999c616b\" (UID: \"a649556e-fd89-4edb-89f7-d1db999c616b\") " Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.724142 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-utilities" (OuterVolumeSpecName: "utilities") pod "a649556e-fd89-4edb-89f7-d1db999c616b" (UID: "a649556e-fd89-4edb-89f7-d1db999c616b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.733800 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a649556e-fd89-4edb-89f7-d1db999c616b-kube-api-access-gjhn7" (OuterVolumeSpecName: "kube-api-access-gjhn7") pod "a649556e-fd89-4edb-89f7-d1db999c616b" (UID: "a649556e-fd89-4edb-89f7-d1db999c616b"). InnerVolumeSpecName "kube-api-access-gjhn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.781389 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a649556e-fd89-4edb-89f7-d1db999c616b" (UID: "a649556e-fd89-4edb-89f7-d1db999c616b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.824644 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjhn7\" (UniqueName: \"kubernetes.io/projected/a649556e-fd89-4edb-89f7-d1db999c616b-kube-api-access-gjhn7\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.824677 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.824690 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a649556e-fd89-4edb-89f7-d1db999c616b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.863224 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50a66a8a-8305-4949-a1c1-09848668b7bc" path="/var/lib/kubelet/pods/50a66a8a-8305-4949-a1c1-09848668b7bc/volumes" Mar 20 16:04:13 crc kubenswrapper[4936]: I0320 16:04:13.864093 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0" path="/var/lib/kubelet/pods/9c4df638-6132-46f8-a8f7-f2b9ef7dc9d0/volumes" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.210647 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" event={"ID":"dacda504-4ad0-4da5-9f22-39cc7a32982b","Type":"ContainerStarted","Data":"c29a75ad24050eece4a5d56bc350d5f98a3d27d1774b8d58afdf3d421316f08e"} Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.212340 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.213861 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" event={"ID":"084dfad4-4a1b-41b8-87fa-693db016d891","Type":"ContainerStarted","Data":"996d079a6eb5501d523f133199fa1b085e9da5377bef5b4ad224be58487cedfa"} Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.213893 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" event={"ID":"084dfad4-4a1b-41b8-87fa-693db016d891","Type":"ContainerStarted","Data":"19f3f6933f581b9b3d4cd3c84a674fd40bbf24b38ab0908c4dbc6bd8d3986a07"} Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.215307 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.218187 4936 generic.go:334] "Generic (PLEG): container finished" podID="a649556e-fd89-4edb-89f7-d1db999c616b" containerID="ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a" exitCode=0 Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.218265 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xp5nm" event={"ID":"a649556e-fd89-4edb-89f7-d1db999c616b","Type":"ContainerDied","Data":"ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a"} Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.218331 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xp5nm" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.218370 4936 scope.go:117] "RemoveContainer" containerID="ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.218346 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xp5nm" event={"ID":"a649556e-fd89-4edb-89f7-d1db999c616b","Type":"ContainerDied","Data":"c475781c82d1343e7b359032f9f5d3316daf4be1f437270b6f129d1fd072b7fc"} Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.222341 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.222877 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.238086 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" podStartSLOduration=4.23806031 podStartE2EDuration="4.23806031s" podCreationTimestamp="2026-03-20 16:04:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:04:14.234301781 +0000 UTC m=+205.180669606" watchObservedRunningTime="2026-03-20 16:04:14.23806031 +0000 UTC m=+205.184428125" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.251770 4936 scope.go:117] "RemoveContainer" containerID="1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.258036 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" podStartSLOduration=4.258011279 podStartE2EDuration="4.258011279s" podCreationTimestamp="2026-03-20 16:04:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:04:14.253621163 +0000 UTC m=+205.199988998" watchObservedRunningTime="2026-03-20 16:04:14.258011279 +0000 UTC m=+205.204379094" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.276818 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xp5nm"] Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.294829 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xp5nm"] Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.299802 4936 scope.go:117] "RemoveContainer" containerID="c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.320101 4936 scope.go:117] "RemoveContainer" containerID="ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a" Mar 20 16:04:14 crc kubenswrapper[4936]: E0320 16:04:14.328271 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a\": container with ID starting with ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a not found: ID does not exist" containerID="ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.328348 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a"} err="failed to get container status \"ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a\": rpc error: code = NotFound desc = could not find container \"ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a\": container with ID starting with ea16da7a156d3b7c01f96bbb3d31b1ba9256582a120f59f9fd0afcb0218e2e5a not found: ID does not exist" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.328395 4936 scope.go:117] "RemoveContainer" containerID="1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0" Mar 20 16:04:14 crc kubenswrapper[4936]: E0320 16:04:14.328865 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0\": container with ID starting with 1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0 not found: ID does not exist" containerID="1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.328895 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0"} err="failed to get container status \"1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0\": rpc error: code = NotFound desc = could not find container \"1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0\": container with ID starting with 1fe52ad4c81f7818a54c8424cbc99f2f7f8bb99708957c965ad16d425ec7acd0 not found: ID does not exist" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.328912 4936 scope.go:117] "RemoveContainer" containerID="c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156" Mar 20 16:04:14 crc kubenswrapper[4936]: E0320 16:04:14.331446 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156\": container with ID starting with c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156 not found: ID does not exist" containerID="c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156" Mar 20 16:04:14 crc kubenswrapper[4936]: I0320 16:04:14.331493 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156"} err="failed to get container status \"c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156\": rpc error: code = NotFound desc = could not find container \"c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156\": container with ID starting with c1cce95c0e7c244f173687e5301b6cfbd04e04155828c2706d83d8e703bcf156 not found: ID does not exist" Mar 20 16:04:15 crc kubenswrapper[4936]: I0320 16:04:15.865172 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" path="/var/lib/kubelet/pods/a649556e-fd89-4edb-89f7-d1db999c616b/volumes" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.248001 4936 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.248993 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" containerName="extract-utilities" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.249017 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" containerName="extract-utilities" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.249038 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" containerName="registry-server" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.249050 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" containerName="registry-server" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.249069 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" containerName="extract-content" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.249081 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" containerName="extract-content" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.249283 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a649556e-fd89-4edb-89f7-d1db999c616b" containerName="registry-server" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.250075 4936 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.250621 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929" gracePeriod=15 Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.250682 4936 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.250721 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316" gracePeriod=15 Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.250800 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8" gracePeriod=15 Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.250773 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749" gracePeriod=15 Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.250640 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62" gracePeriod=15 Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.251182 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.252658 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.252695 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.252719 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.252734 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.252763 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.252779 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.252800 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.252815 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.252854 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.252868 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.252886 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.252899 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.252921 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.252935 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.252959 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.252974 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253213 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253236 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253258 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253282 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253305 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253324 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253346 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253365 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253387 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.253651 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253675 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.253702 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.253720 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.262805 4936 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.305956 4936 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.110:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.354612 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.354649 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.354678 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.354769 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.354802 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.354829 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.354874 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.354906 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456226 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456286 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456330 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456358 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456388 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456427 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456442 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456434 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456471 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456475 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456514 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456586 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456628 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456656 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456721 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.456745 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: I0320 16:04:22.607560 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.640409 4936 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.110:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189e98366618aded openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:04:22.639103469 +0000 UTC m=+213.585471274,LastTimestamp:2026-03-20 16:04:22.639103469 +0000 UTC m=+213.585471274,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:04:22 crc kubenswrapper[4936]: E0320 16:04:22.650728 4936 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.110:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189e98366618aded openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:04:22.639103469 +0000 UTC m=+213.585471274,LastTimestamp:2026-03-20 16:04:22.639103469 +0000 UTC m=+213.585471274,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.286646 4936 generic.go:334] "Generic (PLEG): container finished" podID="87e9a750-951a-40e5-b714-e128787afa82" containerID="fc6d08a42d139f322b64022a1e8003e18664364d819f5a80b9f6a623906e2a31" exitCode=0 Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.286752 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"87e9a750-951a-40e5-b714-e128787afa82","Type":"ContainerDied","Data":"fc6d08a42d139f322b64022a1e8003e18664364d819f5a80b9f6a623906e2a31"} Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.288129 4936 status_manager.go:851] "Failed to get status for pod" podUID="87e9a750-951a-40e5-b714-e128787afa82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.292461 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.294527 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.296358 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62" exitCode=0 Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.296422 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749" exitCode=0 Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.296442 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316" exitCode=0 Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.296461 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8" exitCode=2 Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.296595 4936 scope.go:117] "RemoveContainer" containerID="034e3d72740ad46300abcd522dda0df3dbd1045a617d268c712f30256d833594" Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.299246 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ab0f3b6c0e2e634f7f33fb5c5abc738e8209161a92d4798453a41ff8c2cc19f5"} Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.299311 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a93a42ee95631f75c487d6f7dee97f55c9020a6691ee40a0e43ec1a64f0b8fa5"} Mar 20 16:04:23 crc kubenswrapper[4936]: I0320 16:04:23.300631 4936 status_manager.go:851] "Failed to get status for pod" podUID="87e9a750-951a-40e5-b714-e128787afa82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:23 crc kubenswrapper[4936]: E0320 16:04:23.300655 4936 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.110:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.361851 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.769125 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.769719 4936 status_manager.go:851] "Failed to get status for pod" podUID="87e9a750-951a-40e5-b714-e128787afa82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.791584 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87e9a750-951a-40e5-b714-e128787afa82-kube-api-access\") pod \"87e9a750-951a-40e5-b714-e128787afa82\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.791926 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-var-lock\") pod \"87e9a750-951a-40e5-b714-e128787afa82\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.792023 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-var-lock" (OuterVolumeSpecName: "var-lock") pod "87e9a750-951a-40e5-b714-e128787afa82" (UID: "87e9a750-951a-40e5-b714-e128787afa82"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.792153 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-kubelet-dir\") pod \"87e9a750-951a-40e5-b714-e128787afa82\" (UID: \"87e9a750-951a-40e5-b714-e128787afa82\") " Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.792290 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "87e9a750-951a-40e5-b714-e128787afa82" (UID: "87e9a750-951a-40e5-b714-e128787afa82"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.792627 4936 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-var-lock\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.792697 4936 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87e9a750-951a-40e5-b714-e128787afa82-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.796826 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e9a750-951a-40e5-b714-e128787afa82-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "87e9a750-951a-40e5-b714-e128787afa82" (UID: "87e9a750-951a-40e5-b714-e128787afa82"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:04:24 crc kubenswrapper[4936]: I0320 16:04:24.894499 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87e9a750-951a-40e5-b714-e128787afa82-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.157917 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.159175 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.159942 4936 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.160444 4936 status_manager.go:851] "Failed to get status for pod" podUID="87e9a750-951a-40e5-b714-e128787afa82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.201118 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.201227 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.201271 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.201676 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.201747 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.201750 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.303800 4936 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.304033 4936 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.304579 4936 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.372855 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.373641 4936 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929" exitCode=0 Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.373710 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.373721 4936 scope.go:117] "RemoveContainer" containerID="f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.375641 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"87e9a750-951a-40e5-b714-e128787afa82","Type":"ContainerDied","Data":"1e35e0ff39b17839a6df55182a480f18a282095d8c3bbba77f2c9c2ddd2f513e"} Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.375758 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e35e0ff39b17839a6df55182a480f18a282095d8c3bbba77f2c9c2ddd2f513e" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.375682 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.393453 4936 scope.go:117] "RemoveContainer" containerID="e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.393508 4936 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.394225 4936 status_manager.go:851] "Failed to get status for pod" podUID="87e9a750-951a-40e5-b714-e128787afa82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.398198 4936 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.398492 4936 status_manager.go:851] "Failed to get status for pod" podUID="87e9a750-951a-40e5-b714-e128787afa82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.408892 4936 scope.go:117] "RemoveContainer" containerID="7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.422070 4936 scope.go:117] "RemoveContainer" containerID="45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.437065 4936 scope.go:117] "RemoveContainer" containerID="6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.455598 4936 scope.go:117] "RemoveContainer" containerID="34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.477280 4936 scope.go:117] "RemoveContainer" containerID="f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62" Mar 20 16:04:25 crc kubenswrapper[4936]: E0320 16:04:25.478133 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62\": container with ID starting with f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62 not found: ID does not exist" containerID="f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.478181 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62"} err="failed to get container status \"f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62\": rpc error: code = NotFound desc = could not find container \"f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62\": container with ID starting with f8ff70844dbf765ebe808fbefc8e1e2011a283d50d985d0dd436cf9f5ae26d62 not found: ID does not exist" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.479718 4936 scope.go:117] "RemoveContainer" containerID="e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749" Mar 20 16:04:25 crc kubenswrapper[4936]: E0320 16:04:25.480126 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749\": container with ID starting with e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749 not found: ID does not exist" containerID="e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.480149 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749"} err="failed to get container status \"e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749\": rpc error: code = NotFound desc = could not find container \"e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749\": container with ID starting with e2274fe628c4a2c5f19b65d22797c074f907ac727108180ce14628aa1f8ed749 not found: ID does not exist" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.480162 4936 scope.go:117] "RemoveContainer" containerID="7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316" Mar 20 16:04:25 crc kubenswrapper[4936]: E0320 16:04:25.480621 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316\": container with ID starting with 7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316 not found: ID does not exist" containerID="7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.480669 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316"} err="failed to get container status \"7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316\": rpc error: code = NotFound desc = could not find container \"7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316\": container with ID starting with 7d9e19d7604ec37d801fd7059ada4718da96ef7a09859ab388bca4cef71ef316 not found: ID does not exist" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.480701 4936 scope.go:117] "RemoveContainer" containerID="45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8" Mar 20 16:04:25 crc kubenswrapper[4936]: E0320 16:04:25.480977 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8\": container with ID starting with 45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8 not found: ID does not exist" containerID="45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.481009 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8"} err="failed to get container status \"45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8\": rpc error: code = NotFound desc = could not find container \"45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8\": container with ID starting with 45c5c3127411f3b8f8b3653d14d139377f29eadfb9bc3189fee08a33451fe4b8 not found: ID does not exist" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.481024 4936 scope.go:117] "RemoveContainer" containerID="6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929" Mar 20 16:04:25 crc kubenswrapper[4936]: E0320 16:04:25.481496 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929\": container with ID starting with 6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929 not found: ID does not exist" containerID="6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.481531 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929"} err="failed to get container status \"6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929\": rpc error: code = NotFound desc = could not find container \"6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929\": container with ID starting with 6cc428099d05c84a4ee0f24c06f214c994b7ad4307eef1c3780391415d366929 not found: ID does not exist" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.481562 4936 scope.go:117] "RemoveContainer" containerID="34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133" Mar 20 16:04:25 crc kubenswrapper[4936]: E0320 16:04:25.481808 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133\": container with ID starting with 34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133 not found: ID does not exist" containerID="34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.481827 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133"} err="failed to get container status \"34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133\": rpc error: code = NotFound desc = could not find container \"34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133\": container with ID starting with 34d3ceffa469ff1f882a01874be746533077f1a12691f6a2918b03236b876133 not found: ID does not exist" Mar 20 16:04:25 crc kubenswrapper[4936]: I0320 16:04:25.862870 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 20 16:04:29 crc kubenswrapper[4936]: I0320 16:04:29.859723 4936 status_manager.go:851] "Failed to get status for pod" podUID="87e9a750-951a-40e5-b714-e128787afa82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:30 crc kubenswrapper[4936]: E0320 16:04:30.142794 4936 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:30 crc kubenswrapper[4936]: E0320 16:04:30.143505 4936 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:30 crc kubenswrapper[4936]: E0320 16:04:30.144129 4936 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:30 crc kubenswrapper[4936]: E0320 16:04:30.144433 4936 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:30 crc kubenswrapper[4936]: E0320 16:04:30.144798 4936 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:30 crc kubenswrapper[4936]: I0320 16:04:30.144840 4936 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 20 16:04:30 crc kubenswrapper[4936]: E0320 16:04:30.145142 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="200ms" Mar 20 16:04:30 crc kubenswrapper[4936]: E0320 16:04:30.346253 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="400ms" Mar 20 16:04:30 crc kubenswrapper[4936]: E0320 16:04:30.747878 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="800ms" Mar 20 16:04:31 crc kubenswrapper[4936]: E0320 16:04:31.549309 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="1.6s" Mar 20 16:04:32 crc kubenswrapper[4936]: E0320 16:04:32.652325 4936 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.110:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189e98366618aded openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 16:04:22.639103469 +0000 UTC m=+213.585471274,LastTimestamp:2026-03-20 16:04:22.639103469 +0000 UTC m=+213.585471274,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 16:04:33 crc kubenswrapper[4936]: E0320 16:04:33.151146 4936 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.110:6443: connect: connection refused" interval="3.2s" Mar 20 16:04:33 crc kubenswrapper[4936]: I0320 16:04:33.855899 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:33 crc kubenswrapper[4936]: I0320 16:04:33.859832 4936 status_manager.go:851] "Failed to get status for pod" podUID="87e9a750-951a-40e5-b714-e128787afa82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:33 crc kubenswrapper[4936]: I0320 16:04:33.889788 4936 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:33 crc kubenswrapper[4936]: I0320 16:04:33.889868 4936 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:33 crc kubenswrapper[4936]: E0320 16:04:33.890425 4936 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:33 crc kubenswrapper[4936]: I0320 16:04:33.891091 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:34 crc kubenswrapper[4936]: I0320 16:04:34.443195 4936 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="25057e9353e07e9d121e127cffcad18142fc4c48408ac4fd51df4e87bd9d926e" exitCode=0 Mar 20 16:04:34 crc kubenswrapper[4936]: I0320 16:04:34.443329 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"25057e9353e07e9d121e127cffcad18142fc4c48408ac4fd51df4e87bd9d926e"} Mar 20 16:04:34 crc kubenswrapper[4936]: I0320 16:04:34.443758 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ac27395d32d69d712fcd335d723ec5d32bfceb9fb8241dcd363b029253aa2a39"} Mar 20 16:04:34 crc kubenswrapper[4936]: I0320 16:04:34.444174 4936 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:34 crc kubenswrapper[4936]: I0320 16:04:34.444199 4936 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:34 crc kubenswrapper[4936]: E0320 16:04:34.444855 4936 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:34 crc kubenswrapper[4936]: I0320 16:04:34.444852 4936 status_manager.go:851] "Failed to get status for pod" podUID="87e9a750-951a-40e5-b714-e128787afa82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.110:6443: connect: connection refused" Mar 20 16:04:35 crc kubenswrapper[4936]: I0320 16:04:35.455062 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2d9e33b1c67654db0ebd9c5cbf4aeb3aad4183ef28c2b3df0ba0adb678170737"} Mar 20 16:04:35 crc kubenswrapper[4936]: I0320 16:04:35.455374 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"706168978f890161cbaa85e377b86d6693143b0ae0da14c11a92bc50a666e1af"} Mar 20 16:04:35 crc kubenswrapper[4936]: I0320 16:04:35.455387 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9c712d9454a4bfbe0631b387d73981570554202e0cc6502523442bf2dcc67e4e"} Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.465457 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"28e5e49f8ae92271933b83dce057cc59c34f2042a2d5a6966c6615f25965a8bf"} Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.465554 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"67d9a62c43bc33578095d26f77406e2ac0eafb4823658c602e759cf1f840a0e9"} Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.465974 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.466199 4936 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.466260 4936 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.467642 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.469136 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.469205 4936 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c1efc7a1f0b98ede844eee4ce096aa707924942a4460af4ac095aef8f0c9acc3" exitCode=1 Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.469252 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c1efc7a1f0b98ede844eee4ce096aa707924942a4460af4ac095aef8f0c9acc3"} Mar 20 16:04:36 crc kubenswrapper[4936]: I0320 16:04:36.470131 4936 scope.go:117] "RemoveContainer" containerID="c1efc7a1f0b98ede844eee4ce096aa707924942a4460af4ac095aef8f0c9acc3" Mar 20 16:04:37 crc kubenswrapper[4936]: I0320 16:04:37.292997 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:04:37 crc kubenswrapper[4936]: I0320 16:04:37.478631 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 20 16:04:37 crc kubenswrapper[4936]: I0320 16:04:37.479941 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 16:04:37 crc kubenswrapper[4936]: I0320 16:04:37.479996 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"210763ef522788b72b13e19878b026db578278592e0421a56bd2e46588e47611"} Mar 20 16:04:38 crc kubenswrapper[4936]: I0320 16:04:38.891401 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:38 crc kubenswrapper[4936]: I0320 16:04:38.891462 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:38 crc kubenswrapper[4936]: I0320 16:04:38.901810 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:41 crc kubenswrapper[4936]: I0320 16:04:41.479736 4936 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:41 crc kubenswrapper[4936]: I0320 16:04:41.521012 4936 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:41 crc kubenswrapper[4936]: I0320 16:04:41.521053 4936 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:41 crc kubenswrapper[4936]: I0320 16:04:41.523682 4936 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="601875e3-c976-443b-a73d-3b828affcbab" Mar 20 16:04:41 crc kubenswrapper[4936]: I0320 16:04:41.528888 4936 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://9c712d9454a4bfbe0631b387d73981570554202e0cc6502523442bf2dcc67e4e" Mar 20 16:04:41 crc kubenswrapper[4936]: I0320 16:04:41.528930 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:04:42 crc kubenswrapper[4936]: I0320 16:04:42.529170 4936 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:42 crc kubenswrapper[4936]: I0320 16:04:42.529220 4936 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a183fab9-f13e-47af-a1a5-cc5540d35782" Mar 20 16:04:42 crc kubenswrapper[4936]: I0320 16:04:42.533949 4936 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="601875e3-c976-443b-a73d-3b828affcbab" Mar 20 16:04:43 crc kubenswrapper[4936]: I0320 16:04:43.226959 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:04:43 crc kubenswrapper[4936]: I0320 16:04:43.227189 4936 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 20 16:04:43 crc kubenswrapper[4936]: I0320 16:04:43.227646 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 20 16:04:47 crc kubenswrapper[4936]: I0320 16:04:47.293373 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:04:47 crc kubenswrapper[4936]: I0320 16:04:47.990071 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 16:04:51 crc kubenswrapper[4936]: I0320 16:04:51.273374 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 20 16:04:51 crc kubenswrapper[4936]: I0320 16:04:51.385658 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 20 16:04:52 crc kubenswrapper[4936]: I0320 16:04:52.439069 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 20 16:04:52 crc kubenswrapper[4936]: I0320 16:04:52.634786 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 20 16:04:52 crc kubenswrapper[4936]: I0320 16:04:52.926741 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.097908 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.160792 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.227258 4936 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.227367 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.533658 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.719064 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.721578 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.744600 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.754676 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.873850 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.910149 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 20 16:04:53 crc kubenswrapper[4936]: I0320 16:04:53.946221 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 20 16:04:54 crc kubenswrapper[4936]: I0320 16:04:54.009731 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 20 16:04:54 crc kubenswrapper[4936]: I0320 16:04:54.085894 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 20 16:04:54 crc kubenswrapper[4936]: I0320 16:04:54.128950 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 20 16:04:54 crc kubenswrapper[4936]: I0320 16:04:54.582531 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 20 16:04:54 crc kubenswrapper[4936]: I0320 16:04:54.775429 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 20 16:04:54 crc kubenswrapper[4936]: I0320 16:04:54.820668 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 20 16:04:54 crc kubenswrapper[4936]: I0320 16:04:54.925235 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 20 16:04:54 crc kubenswrapper[4936]: I0320 16:04:54.958280 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 20 16:04:54 crc kubenswrapper[4936]: I0320 16:04:54.967478 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.000734 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.218584 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.233762 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.391340 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.513480 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.631458 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.687736 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.752259 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.759082 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.811438 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.838484 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.864040 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.964724 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 20 16:04:55 crc kubenswrapper[4936]: I0320 16:04:55.982999 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.066930 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.097258 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.149039 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.163199 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.284342 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.287848 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.438988 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.632952 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.639635 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.651103 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.708573 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.747627 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.750483 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.916352 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 20 16:04:56 crc kubenswrapper[4936]: I0320 16:04:56.937460 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.105438 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.106821 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.113679 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.120734 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.168017 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.210986 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.246296 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.248504 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.277827 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.305742 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.400288 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.524160 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.616924 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.647013 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.666413 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.683473 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.799107 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.811572 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 20 16:04:57 crc kubenswrapper[4936]: I0320 16:04:57.953317 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.060682 4936 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.079515 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.090002 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.095262 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.124999 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.125574 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.139524 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.250532 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.252205 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.418067 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.484504 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.551050 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.585335 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.589251 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.603245 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.640825 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.751718 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.764668 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.793437 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.812742 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.836849 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.855764 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.855845 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.872252 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.929340 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.977893 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 16:04:58 crc kubenswrapper[4936]: I0320 16:04:58.978088 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.026823 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.041943 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.263779 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.293873 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.346225 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.422239 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.546627 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.723986 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.741534 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.762269 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.882810 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.920622 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 20 16:04:59 crc kubenswrapper[4936]: I0320 16:04:59.984184 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.044305 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.049280 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.050931 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.067405 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.071341 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.092158 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.098282 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.384947 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.391196 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.413042 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.459209 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.525174 4936 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.533608 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.550329 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.582919 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.631623 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.762805 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.798004 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.963811 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.967845 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 20 16:05:00 crc kubenswrapper[4936]: I0320 16:05:00.993599 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.069394 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.075651 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.076877 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.130390 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.146129 4936 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.212958 4936 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.237231 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.247877 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.386501 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.462585 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.549292 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.565457 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.640981 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.761311 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.807037 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.834089 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.855761 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.864622 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.948183 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 20 16:05:01 crc kubenswrapper[4936]: I0320 16:05:01.978201 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.001089 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.007958 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.014978 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.018971 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.093505 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.157941 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.164874 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.248329 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.376366 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.413048 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.424885 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.438864 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.517074 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.697183 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.773491 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.792354 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.792721 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.819239 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.840006 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.891380 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 20 16:05:02 crc kubenswrapper[4936]: I0320 16:05:02.962639 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.084774 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.097329 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.149425 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.209343 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.227161 4936 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.227246 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.227355 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.228607 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"210763ef522788b72b13e19878b026db578278592e0421a56bd2e46588e47611"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.228848 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://210763ef522788b72b13e19878b026db578278592e0421a56bd2e46588e47611" gracePeriod=30 Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.295082 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.342235 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.349907 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.399371 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.435226 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.478876 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.512769 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.552990 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.584741 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.617936 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.646415 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.698212 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.718688 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.845219 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.909849 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 20 16:05:03 crc kubenswrapper[4936]: I0320 16:05:03.928896 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.027786 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.139105 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.243924 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.489982 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.545961 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.678594 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.751099 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.774240 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.889666 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.896853 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 20 16:05:04 crc kubenswrapper[4936]: I0320 16:05:04.958153 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.031036 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.058829 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.205914 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.447177 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.494706 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.620302 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.689178 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.704840 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.827276 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.835890 4936 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.840936 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.841000 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.848332 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.852178 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.897578 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.901424 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=24.901404678 podStartE2EDuration="24.901404678s" podCreationTimestamp="2026-03-20 16:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:05:05.867678844 +0000 UTC m=+256.814046699" watchObservedRunningTime="2026-03-20 16:05:05.901404678 +0000 UTC m=+256.847772503" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.940692 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.975750 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 20 16:05:05 crc kubenswrapper[4936]: I0320 16:05:05.989777 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.132111 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.378697 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.378811 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.533506 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.626632 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.636281 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.639652 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.660775 4936 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.674173 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.705169 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.731500 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.732637 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 20 16:05:06 crc kubenswrapper[4936]: I0320 16:05:06.987411 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.028319 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.218005 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.225904 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.248297 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.378054 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.488308 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.577167 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.655114 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.664373 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 20 16:05:07 crc kubenswrapper[4936]: I0320 16:05:07.988469 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 20 16:05:08 crc kubenswrapper[4936]: I0320 16:05:08.473833 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 20 16:05:08 crc kubenswrapper[4936]: I0320 16:05:08.504594 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 20 16:05:08 crc kubenswrapper[4936]: I0320 16:05:08.561800 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 20 16:05:08 crc kubenswrapper[4936]: I0320 16:05:08.720075 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 16:05:08 crc kubenswrapper[4936]: I0320 16:05:08.830337 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 20 16:05:08 crc kubenswrapper[4936]: I0320 16:05:08.834174 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 20 16:05:09 crc kubenswrapper[4936]: I0320 16:05:09.434706 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 20 16:05:09 crc kubenswrapper[4936]: I0320 16:05:09.543837 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 20 16:05:09 crc kubenswrapper[4936]: I0320 16:05:09.954598 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 20 16:05:10 crc kubenswrapper[4936]: I0320 16:05:10.605669 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 20 16:05:15 crc kubenswrapper[4936]: I0320 16:05:15.205521 4936 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 16:05:15 crc kubenswrapper[4936]: I0320 16:05:15.206258 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://ab0f3b6c0e2e634f7f33fb5c5abc738e8209161a92d4798453a41ff8c2cc19f5" gracePeriod=5 Mar 20 16:05:20 crc kubenswrapper[4936]: I0320 16:05:20.820817 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 16:05:20 crc kubenswrapper[4936]: I0320 16:05:20.821530 4936 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="ab0f3b6c0e2e634f7f33fb5c5abc738e8209161a92d4798453a41ff8c2cc19f5" exitCode=137 Mar 20 16:05:20 crc kubenswrapper[4936]: I0320 16:05:20.893156 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 16:05:20 crc kubenswrapper[4936]: I0320 16:05:20.893245 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.064226 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.064298 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.064404 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.064472 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.064492 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.064537 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.064608 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.064633 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.064629 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.065428 4936 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.065469 4936 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.065484 4936 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.065495 4936 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.077205 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.166843 4936 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.834708 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.835316 4936 scope.go:117] "RemoveContainer" containerID="ab0f3b6c0e2e634f7f33fb5c5abc738e8209161a92d4798453a41ff8c2cc19f5" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.835428 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 16:05:21 crc kubenswrapper[4936]: I0320 16:05:21.863251 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.877418 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5cpcn"] Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.878253 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5cpcn" podUID="7964eb91-819c-4853-a598-9b93346f496f" containerName="registry-server" containerID="cri-o://110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b" gracePeriod=30 Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.886805 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-km59s"] Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.887279 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-km59s" podUID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerName="registry-server" containerID="cri-o://ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb" gracePeriod=30 Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.904233 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r4w72"] Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.904453 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" podUID="967b1700-c689-4195-8dcd-4177b702d066" containerName="marketplace-operator" containerID="cri-o://7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb" gracePeriod=30 Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.922808 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-njqd9"] Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.923128 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-njqd9" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerName="registry-server" containerID="cri-o://e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6" gracePeriod=30 Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.933856 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kg66h"] Mar 20 16:05:23 crc kubenswrapper[4936]: I0320 16:05:23.934228 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kg66h" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerName="registry-server" containerID="cri-o://8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1" gracePeriod=30 Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.154186 4936 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r4w72 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.154260 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" podUID="967b1700-c689-4195-8dcd-4177b702d066" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.359879 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-km59s" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.416933 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-utilities\") pod \"2aae37a3-f585-4b9b-94e3-5560e06da84c\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.417016 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-catalog-content\") pod \"2aae37a3-f585-4b9b-94e3-5560e06da84c\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.417072 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s5fs\" (UniqueName: \"kubernetes.io/projected/2aae37a3-f585-4b9b-94e3-5560e06da84c-kube-api-access-5s5fs\") pod \"2aae37a3-f585-4b9b-94e3-5560e06da84c\" (UID: \"2aae37a3-f585-4b9b-94e3-5560e06da84c\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.418182 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-utilities" (OuterVolumeSpecName: "utilities") pod "2aae37a3-f585-4b9b-94e3-5560e06da84c" (UID: "2aae37a3-f585-4b9b-94e3-5560e06da84c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.426432 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aae37a3-f585-4b9b-94e3-5560e06da84c-kube-api-access-5s5fs" (OuterVolumeSpecName: "kube-api-access-5s5fs") pod "2aae37a3-f585-4b9b-94e3-5560e06da84c" (UID: "2aae37a3-f585-4b9b-94e3-5560e06da84c"). InnerVolumeSpecName "kube-api-access-5s5fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.491957 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.492793 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2aae37a3-f585-4b9b-94e3-5560e06da84c" (UID: "2aae37a3-f585-4b9b-94e3-5560e06da84c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.499166 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.506401 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.515383 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.517945 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-catalog-content\") pod \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518025 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/967b1700-c689-4195-8dcd-4177b702d066-marketplace-operator-metrics\") pod \"967b1700-c689-4195-8dcd-4177b702d066\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518062 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjjfj\" (UniqueName: \"kubernetes.io/projected/967b1700-c689-4195-8dcd-4177b702d066-kube-api-access-mjjfj\") pod \"967b1700-c689-4195-8dcd-4177b702d066\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518096 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-catalog-content\") pod \"7964eb91-819c-4853-a598-9b93346f496f\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518170 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-utilities\") pod \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518219 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/967b1700-c689-4195-8dcd-4177b702d066-marketplace-trusted-ca\") pod \"967b1700-c689-4195-8dcd-4177b702d066\" (UID: \"967b1700-c689-4195-8dcd-4177b702d066\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518255 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhdlt\" (UniqueName: \"kubernetes.io/projected/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-kube-api-access-nhdlt\") pod \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\" (UID: \"b7e7487c-e0c4-47a9-9875-b1c41e14eecc\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518330 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd9t9\" (UniqueName: \"kubernetes.io/projected/7964eb91-819c-4853-a598-9b93346f496f-kube-api-access-vd9t9\") pod \"7964eb91-819c-4853-a598-9b93346f496f\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518373 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-utilities\") pod \"7964eb91-819c-4853-a598-9b93346f496f\" (UID: \"7964eb91-819c-4853-a598-9b93346f496f\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518415 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj5v4\" (UniqueName: \"kubernetes.io/projected/959c1ced-7501-4754-8e17-c3a420edd1b8-kube-api-access-dj5v4\") pod \"959c1ced-7501-4754-8e17-c3a420edd1b8\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518452 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-utilities\") pod \"959c1ced-7501-4754-8e17-c3a420edd1b8\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518486 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-catalog-content\") pod \"959c1ced-7501-4754-8e17-c3a420edd1b8\" (UID: \"959c1ced-7501-4754-8e17-c3a420edd1b8\") " Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518855 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518892 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aae37a3-f585-4b9b-94e3-5560e06da84c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.518913 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s5fs\" (UniqueName: \"kubernetes.io/projected/2aae37a3-f585-4b9b-94e3-5560e06da84c-kube-api-access-5s5fs\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.519980 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/967b1700-c689-4195-8dcd-4177b702d066-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "967b1700-c689-4195-8dcd-4177b702d066" (UID: "967b1700-c689-4195-8dcd-4177b702d066"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.520325 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-utilities" (OuterVolumeSpecName: "utilities") pod "7964eb91-819c-4853-a598-9b93346f496f" (UID: "7964eb91-819c-4853-a598-9b93346f496f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.521399 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-utilities" (OuterVolumeSpecName: "utilities") pod "b7e7487c-e0c4-47a9-9875-b1c41e14eecc" (UID: "b7e7487c-e0c4-47a9-9875-b1c41e14eecc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.521501 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-utilities" (OuterVolumeSpecName: "utilities") pod "959c1ced-7501-4754-8e17-c3a420edd1b8" (UID: "959c1ced-7501-4754-8e17-c3a420edd1b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.522152 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/967b1700-c689-4195-8dcd-4177b702d066-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "967b1700-c689-4195-8dcd-4177b702d066" (UID: "967b1700-c689-4195-8dcd-4177b702d066"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.522712 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-kube-api-access-nhdlt" (OuterVolumeSpecName: "kube-api-access-nhdlt") pod "b7e7487c-e0c4-47a9-9875-b1c41e14eecc" (UID: "b7e7487c-e0c4-47a9-9875-b1c41e14eecc"). InnerVolumeSpecName "kube-api-access-nhdlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.522858 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/959c1ced-7501-4754-8e17-c3a420edd1b8-kube-api-access-dj5v4" (OuterVolumeSpecName: "kube-api-access-dj5v4") pod "959c1ced-7501-4754-8e17-c3a420edd1b8" (UID: "959c1ced-7501-4754-8e17-c3a420edd1b8"). InnerVolumeSpecName "kube-api-access-dj5v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.523602 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/967b1700-c689-4195-8dcd-4177b702d066-kube-api-access-mjjfj" (OuterVolumeSpecName: "kube-api-access-mjjfj") pod "967b1700-c689-4195-8dcd-4177b702d066" (UID: "967b1700-c689-4195-8dcd-4177b702d066"). InnerVolumeSpecName "kube-api-access-mjjfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.534797 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7964eb91-819c-4853-a598-9b93346f496f-kube-api-access-vd9t9" (OuterVolumeSpecName: "kube-api-access-vd9t9") pod "7964eb91-819c-4853-a598-9b93346f496f" (UID: "7964eb91-819c-4853-a598-9b93346f496f"). InnerVolumeSpecName "kube-api-access-vd9t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.568654 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "959c1ced-7501-4754-8e17-c3a420edd1b8" (UID: "959c1ced-7501-4754-8e17-c3a420edd1b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.594447 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7964eb91-819c-4853-a598-9b93346f496f" (UID: "7964eb91-819c-4853-a598-9b93346f496f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.620435 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd9t9\" (UniqueName: \"kubernetes.io/projected/7964eb91-819c-4853-a598-9b93346f496f-kube-api-access-vd9t9\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.620682 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.620756 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj5v4\" (UniqueName: \"kubernetes.io/projected/959c1ced-7501-4754-8e17-c3a420edd1b8-kube-api-access-dj5v4\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.620820 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.620878 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959c1ced-7501-4754-8e17-c3a420edd1b8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.620939 4936 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/967b1700-c689-4195-8dcd-4177b702d066-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.621004 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjjfj\" (UniqueName: \"kubernetes.io/projected/967b1700-c689-4195-8dcd-4177b702d066-kube-api-access-mjjfj\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.621057 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7964eb91-819c-4853-a598-9b93346f496f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.621116 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.621171 4936 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/967b1700-c689-4195-8dcd-4177b702d066-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.621242 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhdlt\" (UniqueName: \"kubernetes.io/projected/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-kube-api-access-nhdlt\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.677963 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7e7487c-e0c4-47a9-9875-b1c41e14eecc" (UID: "b7e7487c-e0c4-47a9-9875-b1c41e14eecc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.722697 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e7487c-e0c4-47a9-9875-b1c41e14eecc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.861114 4936 generic.go:334] "Generic (PLEG): container finished" podID="7964eb91-819c-4853-a598-9b93346f496f" containerID="110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b" exitCode=0 Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.861193 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5cpcn" event={"ID":"7964eb91-819c-4853-a598-9b93346f496f","Type":"ContainerDied","Data":"110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.861225 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5cpcn" event={"ID":"7964eb91-819c-4853-a598-9b93346f496f","Type":"ContainerDied","Data":"b44f630a17ca56a75599441591c25625a11f2ceaa1a320105c57b3e94b80ba8e"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.861246 4936 scope.go:117] "RemoveContainer" containerID="110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.861368 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5cpcn" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.866472 4936 generic.go:334] "Generic (PLEG): container finished" podID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerID="e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6" exitCode=0 Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.866566 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njqd9" event={"ID":"959c1ced-7501-4754-8e17-c3a420edd1b8","Type":"ContainerDied","Data":"e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.866593 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njqd9" event={"ID":"959c1ced-7501-4754-8e17-c3a420edd1b8","Type":"ContainerDied","Data":"98f458d1fd0de63f1e737aaaadb24d49b86604459029629917df309d5134188d"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.866663 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njqd9" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.876980 4936 generic.go:334] "Generic (PLEG): container finished" podID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerID="ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb" exitCode=0 Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.877121 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km59s" event={"ID":"2aae37a3-f585-4b9b-94e3-5560e06da84c","Type":"ContainerDied","Data":"ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.877158 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-km59s" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.877177 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km59s" event={"ID":"2aae37a3-f585-4b9b-94e3-5560e06da84c","Type":"ContainerDied","Data":"ebd485fa20d242621384574fafa29831784c0ffa71f3c5fcd346fec2e83660a9"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.880893 4936 generic.go:334] "Generic (PLEG): container finished" podID="967b1700-c689-4195-8dcd-4177b702d066" containerID="7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb" exitCode=0 Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.880947 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.881009 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" event={"ID":"967b1700-c689-4195-8dcd-4177b702d066","Type":"ContainerDied","Data":"7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.881047 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r4w72" event={"ID":"967b1700-c689-4195-8dcd-4177b702d066","Type":"ContainerDied","Data":"9ec274b6dffceda00c68ab6638c39b25d361d3bc62936bfae2c51efaf441ed95"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.881230 4936 scope.go:117] "RemoveContainer" containerID="7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.887811 4936 generic.go:334] "Generic (PLEG): container finished" podID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerID="8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1" exitCode=0 Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.887866 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg66h" event={"ID":"b7e7487c-e0c4-47a9-9875-b1c41e14eecc","Type":"ContainerDied","Data":"8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.887908 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg66h" event={"ID":"b7e7487c-e0c4-47a9-9875-b1c41e14eecc","Type":"ContainerDied","Data":"f4d3633c8d428598577285ffd783dca5141f6a3ad903f6323e5ebb24fb49022c"} Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.888005 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg66h" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.910004 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5cpcn"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.934803 4936 scope.go:117] "RemoveContainer" containerID="69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.939881 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5cpcn"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.947670 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-njqd9"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.949718 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-njqd9"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.953472 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r4w72"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.956772 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r4w72"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.966520 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kg66h"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.972590 4936 scope.go:117] "RemoveContainer" containerID="110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b" Mar 20 16:05:24 crc kubenswrapper[4936]: E0320 16:05:24.976908 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b\": container with ID starting with 110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b not found: ID does not exist" containerID="110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.976952 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b"} err="failed to get container status \"110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b\": rpc error: code = NotFound desc = could not find container \"110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b\": container with ID starting with 110524566fd45863ab1bf657af3a22eafb17d14403297e02bb7e606daad7e87b not found: ID does not exist" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.976992 4936 scope.go:117] "RemoveContainer" containerID="7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67" Mar 20 16:05:24 crc kubenswrapper[4936]: E0320 16:05:24.977390 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67\": container with ID starting with 7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67 not found: ID does not exist" containerID="7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.977415 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67"} err="failed to get container status \"7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67\": rpc error: code = NotFound desc = could not find container \"7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67\": container with ID starting with 7161ba41cebd7a67d57b080b36bcd384fce7678c7fba86ffca72721b65d3dc67 not found: ID does not exist" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.977431 4936 scope.go:117] "RemoveContainer" containerID="69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5" Mar 20 16:05:24 crc kubenswrapper[4936]: E0320 16:05:24.977794 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5\": container with ID starting with 69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5 not found: ID does not exist" containerID="69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.977817 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5"} err="failed to get container status \"69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5\": rpc error: code = NotFound desc = could not find container \"69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5\": container with ID starting with 69d6c6c0ecf03ce0a003cc0bd30f8f6078dc40ca5bb4c27c68bf224087f18cd5 not found: ID does not exist" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.977831 4936 scope.go:117] "RemoveContainer" containerID="e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6" Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.981476 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kg66h"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.988177 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-km59s"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.993240 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-km59s"] Mar 20 16:05:24 crc kubenswrapper[4936]: I0320 16:05:24.995518 4936 scope.go:117] "RemoveContainer" containerID="3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.017014 4936 scope.go:117] "RemoveContainer" containerID="ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.061752 4936 scope.go:117] "RemoveContainer" containerID="e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.062573 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6\": container with ID starting with e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6 not found: ID does not exist" containerID="e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.062637 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6"} err="failed to get container status \"e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6\": rpc error: code = NotFound desc = could not find container \"e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6\": container with ID starting with e9893c42a21f83b96208a7c3c6a1ff4b8b329b2190acf8cf7743f232324a60b6 not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.062678 4936 scope.go:117] "RemoveContainer" containerID="3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.063318 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56\": container with ID starting with 3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56 not found: ID does not exist" containerID="3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.063345 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56"} err="failed to get container status \"3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56\": rpc error: code = NotFound desc = could not find container \"3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56\": container with ID starting with 3d58566c1947a3a8bcf4d951d5f34ad4f904b69d45f38d4d2eb9f2fe567b1e56 not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.063366 4936 scope.go:117] "RemoveContainer" containerID="ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.063837 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36\": container with ID starting with ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36 not found: ID does not exist" containerID="ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.063893 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36"} err="failed to get container status \"ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36\": rpc error: code = NotFound desc = could not find container \"ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36\": container with ID starting with ddfb2ce947967e53686d403f1566df033c11723fbd0d540d273352ea4e0a3b36 not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.063923 4936 scope.go:117] "RemoveContainer" containerID="ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.077582 4936 scope.go:117] "RemoveContainer" containerID="6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.095190 4936 scope.go:117] "RemoveContainer" containerID="74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.123276 4936 scope.go:117] "RemoveContainer" containerID="ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.123900 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb\": container with ID starting with ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb not found: ID does not exist" containerID="ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.123968 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb"} err="failed to get container status \"ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb\": rpc error: code = NotFound desc = could not find container \"ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb\": container with ID starting with ca7d8572000f8d2677ec9f8bf7057d8b1e686c712885a5844715401c208338eb not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.124012 4936 scope.go:117] "RemoveContainer" containerID="6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.125264 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de\": container with ID starting with 6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de not found: ID does not exist" containerID="6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.125312 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de"} err="failed to get container status \"6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de\": rpc error: code = NotFound desc = could not find container \"6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de\": container with ID starting with 6579c0b402b867986cc5b3628ad085a60bfc062e86e1886382ecd4430c9b69de not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.125343 4936 scope.go:117] "RemoveContainer" containerID="74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.126051 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22\": container with ID starting with 74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22 not found: ID does not exist" containerID="74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.126163 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22"} err="failed to get container status \"74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22\": rpc error: code = NotFound desc = could not find container \"74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22\": container with ID starting with 74bf523b08365a7d1d9ddbe23b66b4dc18b6f0a07d6f3abb5c5c6c855e388f22 not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.126190 4936 scope.go:117] "RemoveContainer" containerID="7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.141626 4936 scope.go:117] "RemoveContainer" containerID="7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.142230 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb\": container with ID starting with 7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb not found: ID does not exist" containerID="7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.142304 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb"} err="failed to get container status \"7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb\": rpc error: code = NotFound desc = could not find container \"7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb\": container with ID starting with 7736b12c4ec68c3fe7fa47ccf699351e5bf49f64eab7acf8a90151aee9e658fb not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.142362 4936 scope.go:117] "RemoveContainer" containerID="8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.158506 4936 scope.go:117] "RemoveContainer" containerID="3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.182500 4936 scope.go:117] "RemoveContainer" containerID="2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.201874 4936 scope.go:117] "RemoveContainer" containerID="8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.202787 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1\": container with ID starting with 8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1 not found: ID does not exist" containerID="8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.202869 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1"} err="failed to get container status \"8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1\": rpc error: code = NotFound desc = could not find container \"8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1\": container with ID starting with 8f6ec433b91b9104405acddeeb2bf704bcfd357b2a79e222b1ec79bd3c6a17e1 not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.202919 4936 scope.go:117] "RemoveContainer" containerID="3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.203653 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e\": container with ID starting with 3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e not found: ID does not exist" containerID="3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.203719 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e"} err="failed to get container status \"3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e\": rpc error: code = NotFound desc = could not find container \"3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e\": container with ID starting with 3100448955b852cdb07f71d1e82d3047aa7bd794fcffe4e9672e97908a21688e not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.203753 4936 scope.go:117] "RemoveContainer" containerID="2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2" Mar 20 16:05:25 crc kubenswrapper[4936]: E0320 16:05:25.204238 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2\": container with ID starting with 2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2 not found: ID does not exist" containerID="2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.204333 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2"} err="failed to get container status \"2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2\": rpc error: code = NotFound desc = could not find container \"2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2\": container with ID starting with 2d21a82c05e1420741a448e1eb0b95922db69d85547cac2036a815fa96403ef2 not found: ID does not exist" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.864250 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aae37a3-f585-4b9b-94e3-5560e06da84c" path="/var/lib/kubelet/pods/2aae37a3-f585-4b9b-94e3-5560e06da84c/volumes" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.865804 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7964eb91-819c-4853-a598-9b93346f496f" path="/var/lib/kubelet/pods/7964eb91-819c-4853-a598-9b93346f496f/volumes" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.866619 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" path="/var/lib/kubelet/pods/959c1ced-7501-4754-8e17-c3a420edd1b8/volumes" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.868815 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="967b1700-c689-4195-8dcd-4177b702d066" path="/var/lib/kubelet/pods/967b1700-c689-4195-8dcd-4177b702d066/volumes" Mar 20 16:05:25 crc kubenswrapper[4936]: I0320 16:05:25.869487 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" path="/var/lib/kubelet/pods/b7e7487c-e0c4-47a9-9875-b1c41e14eecc/volumes" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758232 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ldm5b"] Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758490 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e9a750-951a-40e5-b714-e128787afa82" containerName="installer" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758505 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e9a750-951a-40e5-b714-e128787afa82" containerName="installer" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758515 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7964eb91-819c-4853-a598-9b93346f496f" containerName="extract-content" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758521 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7964eb91-819c-4853-a598-9b93346f496f" containerName="extract-content" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758530 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7964eb91-819c-4853-a598-9b93346f496f" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758537 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7964eb91-819c-4853-a598-9b93346f496f" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758565 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7964eb91-819c-4853-a598-9b93346f496f" containerName="extract-utilities" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758571 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7964eb91-819c-4853-a598-9b93346f496f" containerName="extract-utilities" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758583 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758589 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758598 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967b1700-c689-4195-8dcd-4177b702d066" containerName="marketplace-operator" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758604 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="967b1700-c689-4195-8dcd-4177b702d066" containerName="marketplace-operator" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758613 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerName="extract-utilities" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758621 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerName="extract-utilities" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758627 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758633 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758642 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerName="extract-content" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758649 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerName="extract-content" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758658 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerName="extract-content" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758665 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerName="extract-content" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758672 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerName="extract-utilities" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758678 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerName="extract-utilities" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758685 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerName="extract-utilities" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758691 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerName="extract-utilities" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758703 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758709 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758716 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758722 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 16:05:28 crc kubenswrapper[4936]: E0320 16:05:28.758734 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerName="extract-content" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758740 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerName="extract-content" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758845 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758859 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e7487c-e0c4-47a9-9875-b1c41e14eecc" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758868 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aae37a3-f585-4b9b-94e3-5560e06da84c" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758878 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e9a750-951a-40e5-b714-e128787afa82" containerName="installer" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758885 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="959c1ced-7501-4754-8e17-c3a420edd1b8" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758894 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="967b1700-c689-4195-8dcd-4177b702d066" containerName="marketplace-operator" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.758900 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="7964eb91-819c-4853-a598-9b93346f496f" containerName="registry-server" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.759322 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.762683 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.763423 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.763588 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.763701 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.770866 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ldm5b"] Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.776577 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.783626 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt"] Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.783884 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" podUID="084dfad4-4a1b-41b8-87fa-693db016d891" containerName="route-controller-manager" containerID="cri-o://996d079a6eb5501d523f133199fa1b085e9da5377bef5b4ad224be58487cedfa" gracePeriod=30 Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.786322 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7977b8c78-h7dxg"] Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.786832 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" podUID="dacda504-4ad0-4da5-9f22-39cc7a32982b" containerName="controller-manager" containerID="cri-o://c29a75ad24050eece4a5d56bc350d5f98a3d27d1774b8d58afdf3d421316f08e" gracePeriod=30 Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.793476 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d82b61cf-0c16-48b7-9a4f-c42b9212d85a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ldm5b\" (UID: \"d82b61cf-0c16-48b7-9a4f-c42b9212d85a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.793527 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsgvz\" (UniqueName: \"kubernetes.io/projected/d82b61cf-0c16-48b7-9a4f-c42b9212d85a-kube-api-access-rsgvz\") pod \"marketplace-operator-79b997595-ldm5b\" (UID: \"d82b61cf-0c16-48b7-9a4f-c42b9212d85a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.793643 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d82b61cf-0c16-48b7-9a4f-c42b9212d85a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ldm5b\" (UID: \"d82b61cf-0c16-48b7-9a4f-c42b9212d85a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.855033 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.855319 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.894502 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d82b61cf-0c16-48b7-9a4f-c42b9212d85a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ldm5b\" (UID: \"d82b61cf-0c16-48b7-9a4f-c42b9212d85a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.894571 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d82b61cf-0c16-48b7-9a4f-c42b9212d85a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ldm5b\" (UID: \"d82b61cf-0c16-48b7-9a4f-c42b9212d85a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.894612 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsgvz\" (UniqueName: \"kubernetes.io/projected/d82b61cf-0c16-48b7-9a4f-c42b9212d85a-kube-api-access-rsgvz\") pod \"marketplace-operator-79b997595-ldm5b\" (UID: \"d82b61cf-0c16-48b7-9a4f-c42b9212d85a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.896528 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d82b61cf-0c16-48b7-9a4f-c42b9212d85a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ldm5b\" (UID: \"d82b61cf-0c16-48b7-9a4f-c42b9212d85a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.906119 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d82b61cf-0c16-48b7-9a4f-c42b9212d85a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ldm5b\" (UID: \"d82b61cf-0c16-48b7-9a4f-c42b9212d85a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.920076 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsgvz\" (UniqueName: \"kubernetes.io/projected/d82b61cf-0c16-48b7-9a4f-c42b9212d85a-kube-api-access-rsgvz\") pod \"marketplace-operator-79b997595-ldm5b\" (UID: \"d82b61cf-0c16-48b7-9a4f-c42b9212d85a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.944332 4936 generic.go:334] "Generic (PLEG): container finished" podID="dacda504-4ad0-4da5-9f22-39cc7a32982b" containerID="c29a75ad24050eece4a5d56bc350d5f98a3d27d1774b8d58afdf3d421316f08e" exitCode=0 Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.944425 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" event={"ID":"dacda504-4ad0-4da5-9f22-39cc7a32982b","Type":"ContainerDied","Data":"c29a75ad24050eece4a5d56bc350d5f98a3d27d1774b8d58afdf3d421316f08e"} Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.947899 4936 generic.go:334] "Generic (PLEG): container finished" podID="084dfad4-4a1b-41b8-87fa-693db016d891" containerID="996d079a6eb5501d523f133199fa1b085e9da5377bef5b4ad224be58487cedfa" exitCode=0 Mar 20 16:05:28 crc kubenswrapper[4936]: I0320 16:05:28.947936 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" event={"ID":"084dfad4-4a1b-41b8-87fa-693db016d891","Type":"ContainerDied","Data":"996d079a6eb5501d523f133199fa1b085e9da5377bef5b4ad224be58487cedfa"} Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.075530 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.317182 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.319015 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.452235 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ldm5b"] Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.503056 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-config\") pod \"dacda504-4ad0-4da5-9f22-39cc7a32982b\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.503100 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/084dfad4-4a1b-41b8-87fa-693db016d891-serving-cert\") pod \"084dfad4-4a1b-41b8-87fa-693db016d891\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.503127 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-proxy-ca-bundles\") pod \"dacda504-4ad0-4da5-9f22-39cc7a32982b\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.503145 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-client-ca\") pod \"084dfad4-4a1b-41b8-87fa-693db016d891\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.503229 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vhrb\" (UniqueName: \"kubernetes.io/projected/084dfad4-4a1b-41b8-87fa-693db016d891-kube-api-access-5vhrb\") pod \"084dfad4-4a1b-41b8-87fa-693db016d891\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.503262 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-config\") pod \"084dfad4-4a1b-41b8-87fa-693db016d891\" (UID: \"084dfad4-4a1b-41b8-87fa-693db016d891\") " Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.503283 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-client-ca\") pod \"dacda504-4ad0-4da5-9f22-39cc7a32982b\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.503309 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lspvq\" (UniqueName: \"kubernetes.io/projected/dacda504-4ad0-4da5-9f22-39cc7a32982b-kube-api-access-lspvq\") pod \"dacda504-4ad0-4da5-9f22-39cc7a32982b\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.503325 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacda504-4ad0-4da5-9f22-39cc7a32982b-serving-cert\") pod \"dacda504-4ad0-4da5-9f22-39cc7a32982b\" (UID: \"dacda504-4ad0-4da5-9f22-39cc7a32982b\") " Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.504105 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-client-ca" (OuterVolumeSpecName: "client-ca") pod "084dfad4-4a1b-41b8-87fa-693db016d891" (UID: "084dfad4-4a1b-41b8-87fa-693db016d891"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.504634 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "dacda504-4ad0-4da5-9f22-39cc7a32982b" (UID: "dacda504-4ad0-4da5-9f22-39cc7a32982b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.504669 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-client-ca" (OuterVolumeSpecName: "client-ca") pod "dacda504-4ad0-4da5-9f22-39cc7a32982b" (UID: "dacda504-4ad0-4da5-9f22-39cc7a32982b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.504833 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-config" (OuterVolumeSpecName: "config") pod "084dfad4-4a1b-41b8-87fa-693db016d891" (UID: "084dfad4-4a1b-41b8-87fa-693db016d891"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.505038 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-config" (OuterVolumeSpecName: "config") pod "dacda504-4ad0-4da5-9f22-39cc7a32982b" (UID: "dacda504-4ad0-4da5-9f22-39cc7a32982b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.509562 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dacda504-4ad0-4da5-9f22-39cc7a32982b-kube-api-access-lspvq" (OuterVolumeSpecName: "kube-api-access-lspvq") pod "dacda504-4ad0-4da5-9f22-39cc7a32982b" (UID: "dacda504-4ad0-4da5-9f22-39cc7a32982b"). InnerVolumeSpecName "kube-api-access-lspvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.510051 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacda504-4ad0-4da5-9f22-39cc7a32982b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dacda504-4ad0-4da5-9f22-39cc7a32982b" (UID: "dacda504-4ad0-4da5-9f22-39cc7a32982b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.512818 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/084dfad4-4a1b-41b8-87fa-693db016d891-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "084dfad4-4a1b-41b8-87fa-693db016d891" (UID: "084dfad4-4a1b-41b8-87fa-693db016d891"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.514106 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/084dfad4-4a1b-41b8-87fa-693db016d891-kube-api-access-5vhrb" (OuterVolumeSpecName: "kube-api-access-5vhrb") pod "084dfad4-4a1b-41b8-87fa-693db016d891" (UID: "084dfad4-4a1b-41b8-87fa-693db016d891"). InnerVolumeSpecName "kube-api-access-5vhrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.608975 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.609020 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.609031 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacda504-4ad0-4da5-9f22-39cc7a32982b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.609042 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lspvq\" (UniqueName: \"kubernetes.io/projected/dacda504-4ad0-4da5-9f22-39cc7a32982b-kube-api-access-lspvq\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.609061 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.609070 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/084dfad4-4a1b-41b8-87fa-693db016d891-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.609079 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dacda504-4ad0-4da5-9f22-39cc7a32982b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.609089 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/084dfad4-4a1b-41b8-87fa-693db016d891-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.609102 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vhrb\" (UniqueName: \"kubernetes.io/projected/084dfad4-4a1b-41b8-87fa-693db016d891-kube-api-access-5vhrb\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.954950 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" event={"ID":"084dfad4-4a1b-41b8-87fa-693db016d891","Type":"ContainerDied","Data":"19f3f6933f581b9b3d4cd3c84a674fd40bbf24b38ab0908c4dbc6bd8d3986a07"} Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.955033 4936 scope.go:117] "RemoveContainer" containerID="996d079a6eb5501d523f133199fa1b085e9da5377bef5b4ad224be58487cedfa" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.955129 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.958619 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" event={"ID":"dacda504-4ad0-4da5-9f22-39cc7a32982b","Type":"ContainerDied","Data":"2b24b331aaf2b35fa684308804d426b732f5059bd7ccb86211e7210bba283d21"} Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.959114 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7977b8c78-h7dxg" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.965954 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" event={"ID":"d82b61cf-0c16-48b7-9a4f-c42b9212d85a","Type":"ContainerStarted","Data":"084d4ff5cc6a2f711601094d49eee5a6d2738df9ac77ada0943d53073bda5d95"} Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.966022 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" event={"ID":"d82b61cf-0c16-48b7-9a4f-c42b9212d85a","Type":"ContainerStarted","Data":"9239e756db8017a53f637449530451df37fa6f95f70e33efc44588bd34788e9b"} Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.969748 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.975792 4936 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ldm5b container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.65:8080/healthz\": dial tcp 10.217.0.65:8080: connect: connection refused" start-of-body= Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.977513 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" podUID="d82b61cf-0c16-48b7-9a4f-c42b9212d85a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.65:8080/healthz\": dial tcp 10.217.0.65:8080: connect: connection refused" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.986116 4936 scope.go:117] "RemoveContainer" containerID="c29a75ad24050eece4a5d56bc350d5f98a3d27d1774b8d58afdf3d421316f08e" Mar 20 16:05:29 crc kubenswrapper[4936]: I0320 16:05:29.994980 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7977b8c78-h7dxg"] Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.004395 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7977b8c78-h7dxg"] Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.011588 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt"] Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.015474 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bf99c7bf-zqszt"] Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.024080 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" podStartSLOduration=2.024043572 podStartE2EDuration="2.024043572s" podCreationTimestamp="2026-03-20 16:05:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:05:30.017134388 +0000 UTC m=+280.963502203" watchObservedRunningTime="2026-03-20 16:05:30.024043572 +0000 UTC m=+280.970411387" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.465528 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh"] Mar 20 16:05:30 crc kubenswrapper[4936]: E0320 16:05:30.466070 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="084dfad4-4a1b-41b8-87fa-693db016d891" containerName="route-controller-manager" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.466084 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="084dfad4-4a1b-41b8-87fa-693db016d891" containerName="route-controller-manager" Mar 20 16:05:30 crc kubenswrapper[4936]: E0320 16:05:30.466095 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dacda504-4ad0-4da5-9f22-39cc7a32982b" containerName="controller-manager" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.466101 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="dacda504-4ad0-4da5-9f22-39cc7a32982b" containerName="controller-manager" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.466193 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="084dfad4-4a1b-41b8-87fa-693db016d891" containerName="route-controller-manager" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.466208 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="dacda504-4ad0-4da5-9f22-39cc7a32982b" containerName="controller-manager" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.466634 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.469404 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.470425 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.470844 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.470897 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.471147 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.473153 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.478830 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74b956d8c5-5jg5j"] Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.479801 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.482339 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.482985 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.483189 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.483533 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.484344 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.486685 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.491620 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.491615 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh"] Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.509642 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74b956d8c5-5jg5j"] Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.521834 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tn2l\" (UniqueName: \"kubernetes.io/projected/93c16989-3ce2-4cd0-87d7-b980254194d5-kube-api-access-6tn2l\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.521907 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93c16989-3ce2-4cd0-87d7-b980254194d5-serving-cert\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.521946 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/231e626c-abc5-4120-9cef-3776b26207ac-serving-cert\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.521977 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-client-ca\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.522011 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-config\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.522067 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-proxy-ca-bundles\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.522142 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-config\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.522171 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-client-ca\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.522205 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhxrf\" (UniqueName: \"kubernetes.io/projected/231e626c-abc5-4120-9cef-3776b26207ac-kube-api-access-jhxrf\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.532981 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74b956d8c5-5jg5j"] Mar 20 16:05:30 crc kubenswrapper[4936]: E0320 16:05:30.533448 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-6tn2l proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" podUID="93c16989-3ce2-4cd0-87d7-b980254194d5" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.566272 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh"] Mar 20 16:05:30 crc kubenswrapper[4936]: E0320 16:05:30.566889 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-jhxrf serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" podUID="231e626c-abc5-4120-9cef-3776b26207ac" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.623975 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-proxy-ca-bundles\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.624042 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-client-ca\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.624063 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-config\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.624089 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhxrf\" (UniqueName: \"kubernetes.io/projected/231e626c-abc5-4120-9cef-3776b26207ac-kube-api-access-jhxrf\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.624123 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tn2l\" (UniqueName: \"kubernetes.io/projected/93c16989-3ce2-4cd0-87d7-b980254194d5-kube-api-access-6tn2l\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.624146 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93c16989-3ce2-4cd0-87d7-b980254194d5-serving-cert\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.624165 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/231e626c-abc5-4120-9cef-3776b26207ac-serving-cert\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.624190 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-client-ca\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.624210 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-config\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.625655 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-config\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.625689 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-client-ca\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.626245 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-client-ca\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.628323 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-config\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.628452 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-proxy-ca-bundles\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.630973 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93c16989-3ce2-4cd0-87d7-b980254194d5-serving-cert\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.636718 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/231e626c-abc5-4120-9cef-3776b26207ac-serving-cert\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.642412 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tn2l\" (UniqueName: \"kubernetes.io/projected/93c16989-3ce2-4cd0-87d7-b980254194d5-kube-api-access-6tn2l\") pod \"controller-manager-74b956d8c5-5jg5j\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.642606 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhxrf\" (UniqueName: \"kubernetes.io/projected/231e626c-abc5-4120-9cef-3776b26207ac-kube-api-access-jhxrf\") pod \"route-controller-manager-59577dcc5d-l5jqh\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.972912 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.972928 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.980653 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ldm5b" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.984059 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:30 crc kubenswrapper[4936]: I0320 16:05:30.990663 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.028227 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tn2l\" (UniqueName: \"kubernetes.io/projected/93c16989-3ce2-4cd0-87d7-b980254194d5-kube-api-access-6tn2l\") pod \"93c16989-3ce2-4cd0-87d7-b980254194d5\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.028334 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-client-ca\") pod \"231e626c-abc5-4120-9cef-3776b26207ac\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.028407 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-config\") pod \"93c16989-3ce2-4cd0-87d7-b980254194d5\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.028439 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-proxy-ca-bundles\") pod \"93c16989-3ce2-4cd0-87d7-b980254194d5\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.028461 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-client-ca\") pod \"93c16989-3ce2-4cd0-87d7-b980254194d5\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.028509 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/231e626c-abc5-4120-9cef-3776b26207ac-serving-cert\") pod \"231e626c-abc5-4120-9cef-3776b26207ac\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.028609 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93c16989-3ce2-4cd0-87d7-b980254194d5-serving-cert\") pod \"93c16989-3ce2-4cd0-87d7-b980254194d5\" (UID: \"93c16989-3ce2-4cd0-87d7-b980254194d5\") " Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.028654 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhxrf\" (UniqueName: \"kubernetes.io/projected/231e626c-abc5-4120-9cef-3776b26207ac-kube-api-access-jhxrf\") pod \"231e626c-abc5-4120-9cef-3776b26207ac\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.028703 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-config\") pod \"231e626c-abc5-4120-9cef-3776b26207ac\" (UID: \"231e626c-abc5-4120-9cef-3776b26207ac\") " Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.029265 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-config" (OuterVolumeSpecName: "config") pod "93c16989-3ce2-4cd0-87d7-b980254194d5" (UID: "93c16989-3ce2-4cd0-87d7-b980254194d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.029493 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "93c16989-3ce2-4cd0-87d7-b980254194d5" (UID: "93c16989-3ce2-4cd0-87d7-b980254194d5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.029628 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-config" (OuterVolumeSpecName: "config") pod "231e626c-abc5-4120-9cef-3776b26207ac" (UID: "231e626c-abc5-4120-9cef-3776b26207ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.029785 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-client-ca" (OuterVolumeSpecName: "client-ca") pod "93c16989-3ce2-4cd0-87d7-b980254194d5" (UID: "93c16989-3ce2-4cd0-87d7-b980254194d5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.029843 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-client-ca" (OuterVolumeSpecName: "client-ca") pod "231e626c-abc5-4120-9cef-3776b26207ac" (UID: "231e626c-abc5-4120-9cef-3776b26207ac"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.033982 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c16989-3ce2-4cd0-87d7-b980254194d5-kube-api-access-6tn2l" (OuterVolumeSpecName: "kube-api-access-6tn2l") pod "93c16989-3ce2-4cd0-87d7-b980254194d5" (UID: "93c16989-3ce2-4cd0-87d7-b980254194d5"). InnerVolumeSpecName "kube-api-access-6tn2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.035525 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/231e626c-abc5-4120-9cef-3776b26207ac-kube-api-access-jhxrf" (OuterVolumeSpecName: "kube-api-access-jhxrf") pod "231e626c-abc5-4120-9cef-3776b26207ac" (UID: "231e626c-abc5-4120-9cef-3776b26207ac"). InnerVolumeSpecName "kube-api-access-jhxrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.036076 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c16989-3ce2-4cd0-87d7-b980254194d5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "93c16989-3ce2-4cd0-87d7-b980254194d5" (UID: "93c16989-3ce2-4cd0-87d7-b980254194d5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.036229 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/231e626c-abc5-4120-9cef-3776b26207ac-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "231e626c-abc5-4120-9cef-3776b26207ac" (UID: "231e626c-abc5-4120-9cef-3776b26207ac"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.130378 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tn2l\" (UniqueName: \"kubernetes.io/projected/93c16989-3ce2-4cd0-87d7-b980254194d5-kube-api-access-6tn2l\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.130438 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.130459 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.130472 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.130487 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93c16989-3ce2-4cd0-87d7-b980254194d5-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.130500 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/231e626c-abc5-4120-9cef-3776b26207ac-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.130516 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93c16989-3ce2-4cd0-87d7-b980254194d5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.130529 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhxrf\" (UniqueName: \"kubernetes.io/projected/231e626c-abc5-4120-9cef-3776b26207ac-kube-api-access-jhxrf\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.130560 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/231e626c-abc5-4120-9cef-3776b26207ac-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.862985 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="084dfad4-4a1b-41b8-87fa-693db016d891" path="/var/lib/kubelet/pods/084dfad4-4a1b-41b8-87fa-693db016d891/volumes" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.863928 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dacda504-4ad0-4da5-9f22-39cc7a32982b" path="/var/lib/kubelet/pods/dacda504-4ad0-4da5-9f22-39cc7a32982b/volumes" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.981518 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh" Mar 20 16:05:31 crc kubenswrapper[4936]: I0320 16:05:31.982373 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74b956d8c5-5jg5j" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.092532 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq"] Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.094289 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.096848 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh"] Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.103107 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.103198 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.103423 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.103499 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.103431 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59577dcc5d-l5jqh"] Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.104212 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.104386 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.122194 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq"] Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.136674 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74b956d8c5-5jg5j"] Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.143764 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74b956d8c5-5jg5j"] Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.246648 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77a37156-be74-4f60-ad0d-9248408b80d4-serving-cert\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.247064 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsrzh\" (UniqueName: \"kubernetes.io/projected/77a37156-be74-4f60-ad0d-9248408b80d4-kube-api-access-gsrzh\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.247094 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-client-ca\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.247131 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-config\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.348320 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77a37156-be74-4f60-ad0d-9248408b80d4-serving-cert\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.349454 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsrzh\" (UniqueName: \"kubernetes.io/projected/77a37156-be74-4f60-ad0d-9248408b80d4-kube-api-access-gsrzh\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.349498 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-client-ca\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.349629 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-config\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.350978 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-config\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.351047 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-client-ca\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.358196 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77a37156-be74-4f60-ad0d-9248408b80d4-serving-cert\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.367901 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsrzh\" (UniqueName: \"kubernetes.io/projected/77a37156-be74-4f60-ad0d-9248408b80d4-kube-api-access-gsrzh\") pod \"route-controller-manager-7ff7586b44-mgrcq\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.429516 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.704020 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq"] Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.991312 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" event={"ID":"77a37156-be74-4f60-ad0d-9248408b80d4","Type":"ContainerStarted","Data":"19ab13ddef3b485fe6601afb704083734d139b6c1aa2a27568617af553a0e75b"} Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.992440 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" event={"ID":"77a37156-be74-4f60-ad0d-9248408b80d4","Type":"ContainerStarted","Data":"a6cf2ae5df35959c21646b48ac56ce371052b14984d6887cc6919dd004291fd4"} Mar 20 16:05:32 crc kubenswrapper[4936]: I0320 16:05:32.992502 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:33 crc kubenswrapper[4936]: I0320 16:05:33.009960 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" podStartSLOduration=3.009937176 podStartE2EDuration="3.009937176s" podCreationTimestamp="2026-03-20 16:05:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:05:33.007842441 +0000 UTC m=+283.954210296" watchObservedRunningTime="2026-03-20 16:05:33.009937176 +0000 UTC m=+283.956305031" Mar 20 16:05:33 crc kubenswrapper[4936]: I0320 16:05:33.576396 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:33 crc kubenswrapper[4936]: I0320 16:05:33.862247 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="231e626c-abc5-4120-9cef-3776b26207ac" path="/var/lib/kubelet/pods/231e626c-abc5-4120-9cef-3776b26207ac/volumes" Mar 20 16:05:33 crc kubenswrapper[4936]: I0320 16:05:33.864800 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93c16989-3ce2-4cd0-87d7-b980254194d5" path="/var/lib/kubelet/pods/93c16989-3ce2-4cd0-87d7-b980254194d5/volumes" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:33.999980 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.001056 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.002518 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.002582 4936 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="210763ef522788b72b13e19878b026db578278592e0421a56bd2e46588e47611" exitCode=137 Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.002771 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"210763ef522788b72b13e19878b026db578278592e0421a56bd2e46588e47611"} Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.002865 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"49efb0644635dc8a2d19306e10b78bb2b8fbdc8cb85f1465fc930f692f4f2d2c"} Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.002888 4936 scope.go:117] "RemoveContainer" containerID="c1efc7a1f0b98ede844eee4ce096aa707924942a4460af4ac095aef8f0c9acc3" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.468459 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-scqrx"] Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.469103 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.471755 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.471826 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.473783 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.473893 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.475029 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.486844 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-scqrx"] Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.495117 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.499227 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.583488 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltsz2\" (UniqueName: \"kubernetes.io/projected/d25b364f-96c6-4513-9142-c198d3a2cbae-kube-api-access-ltsz2\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.583736 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d25b364f-96c6-4513-9142-c198d3a2cbae-serving-cert\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.583806 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-client-ca\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.583866 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-proxy-ca-bundles\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.583922 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-config\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.687370 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltsz2\" (UniqueName: \"kubernetes.io/projected/d25b364f-96c6-4513-9142-c198d3a2cbae-kube-api-access-ltsz2\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.687612 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d25b364f-96c6-4513-9142-c198d3a2cbae-serving-cert\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.687673 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-client-ca\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.687727 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-proxy-ca-bundles\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.687787 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-config\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.690645 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-client-ca\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.691506 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-proxy-ca-bundles\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.692763 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-config\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.696144 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d25b364f-96c6-4513-9142-c198d3a2cbae-serving-cert\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.725337 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltsz2\" (UniqueName: \"kubernetes.io/projected/d25b364f-96c6-4513-9142-c198d3a2cbae-kube-api-access-ltsz2\") pod \"controller-manager-5c8985447-scqrx\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:34 crc kubenswrapper[4936]: I0320 16:05:34.809177 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:35 crc kubenswrapper[4936]: I0320 16:05:35.014911 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 20 16:05:35 crc kubenswrapper[4936]: I0320 16:05:35.016159 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 20 16:05:35 crc kubenswrapper[4936]: I0320 16:05:35.280269 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-scqrx"] Mar 20 16:05:36 crc kubenswrapper[4936]: I0320 16:05:36.026134 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" event={"ID":"d25b364f-96c6-4513-9142-c198d3a2cbae","Type":"ContainerStarted","Data":"8728a72624f65e72cce1840b657ecc386792f304a89b0195276a6f7cc1ca7173"} Mar 20 16:05:36 crc kubenswrapper[4936]: I0320 16:05:36.026199 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" event={"ID":"d25b364f-96c6-4513-9142-c198d3a2cbae","Type":"ContainerStarted","Data":"7d99ba61184d77005a476969e3aa4eaaf6103c9f93bdfec53f96e4324099123d"} Mar 20 16:05:36 crc kubenswrapper[4936]: I0320 16:05:36.026420 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:36 crc kubenswrapper[4936]: I0320 16:05:36.031278 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:36 crc kubenswrapper[4936]: I0320 16:05:36.056134 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" podStartSLOduration=6.05610503 podStartE2EDuration="6.05610503s" podCreationTimestamp="2026-03-20 16:05:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:05:36.051796816 +0000 UTC m=+286.998164641" watchObservedRunningTime="2026-03-20 16:05:36.05610503 +0000 UTC m=+287.002472845" Mar 20 16:05:37 crc kubenswrapper[4936]: I0320 16:05:37.293370 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:05:43 crc kubenswrapper[4936]: I0320 16:05:43.226646 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:05:43 crc kubenswrapper[4936]: I0320 16:05:43.230318 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:05:44 crc kubenswrapper[4936]: I0320 16:05:44.087989 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 16:05:52 crc kubenswrapper[4936]: I0320 16:05:52.807328 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq"] Mar 20 16:05:52 crc kubenswrapper[4936]: I0320 16:05:52.807956 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" podUID="77a37156-be74-4f60-ad0d-9248408b80d4" containerName="route-controller-manager" containerID="cri-o://19ab13ddef3b485fe6601afb704083734d139b6c1aa2a27568617af553a0e75b" gracePeriod=30 Mar 20 16:05:52 crc kubenswrapper[4936]: I0320 16:05:52.843225 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-scqrx"] Mar 20 16:05:52 crc kubenswrapper[4936]: I0320 16:05:52.843451 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" podUID="d25b364f-96c6-4513-9142-c198d3a2cbae" containerName="controller-manager" containerID="cri-o://8728a72624f65e72cce1840b657ecc386792f304a89b0195276a6f7cc1ca7173" gracePeriod=30 Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.136092 4936 generic.go:334] "Generic (PLEG): container finished" podID="77a37156-be74-4f60-ad0d-9248408b80d4" containerID="19ab13ddef3b485fe6601afb704083734d139b6c1aa2a27568617af553a0e75b" exitCode=0 Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.136215 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" event={"ID":"77a37156-be74-4f60-ad0d-9248408b80d4","Type":"ContainerDied","Data":"19ab13ddef3b485fe6601afb704083734d139b6c1aa2a27568617af553a0e75b"} Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.138776 4936 generic.go:334] "Generic (PLEG): container finished" podID="d25b364f-96c6-4513-9142-c198d3a2cbae" containerID="8728a72624f65e72cce1840b657ecc386792f304a89b0195276a6f7cc1ca7173" exitCode=0 Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.138807 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" event={"ID":"d25b364f-96c6-4513-9142-c198d3a2cbae","Type":"ContainerDied","Data":"8728a72624f65e72cce1840b657ecc386792f304a89b0195276a6f7cc1ca7173"} Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.396146 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.473285 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77a37156-be74-4f60-ad0d-9248408b80d4-serving-cert\") pod \"77a37156-be74-4f60-ad0d-9248408b80d4\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.473456 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-config\") pod \"77a37156-be74-4f60-ad0d-9248408b80d4\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.473507 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsrzh\" (UniqueName: \"kubernetes.io/projected/77a37156-be74-4f60-ad0d-9248408b80d4-kube-api-access-gsrzh\") pod \"77a37156-be74-4f60-ad0d-9248408b80d4\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.473558 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-client-ca\") pod \"77a37156-be74-4f60-ad0d-9248408b80d4\" (UID: \"77a37156-be74-4f60-ad0d-9248408b80d4\") " Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.474704 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-client-ca" (OuterVolumeSpecName: "client-ca") pod "77a37156-be74-4f60-ad0d-9248408b80d4" (UID: "77a37156-be74-4f60-ad0d-9248408b80d4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.475670 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-config" (OuterVolumeSpecName: "config") pod "77a37156-be74-4f60-ad0d-9248408b80d4" (UID: "77a37156-be74-4f60-ad0d-9248408b80d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.481302 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.482513 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77a37156-be74-4f60-ad0d-9248408b80d4-kube-api-access-gsrzh" (OuterVolumeSpecName: "kube-api-access-gsrzh") pod "77a37156-be74-4f60-ad0d-9248408b80d4" (UID: "77a37156-be74-4f60-ad0d-9248408b80d4"). InnerVolumeSpecName "kube-api-access-gsrzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.483009 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77a37156-be74-4f60-ad0d-9248408b80d4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "77a37156-be74-4f60-ad0d-9248408b80d4" (UID: "77a37156-be74-4f60-ad0d-9248408b80d4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.574815 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-proxy-ca-bundles\") pod \"d25b364f-96c6-4513-9142-c198d3a2cbae\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.574887 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltsz2\" (UniqueName: \"kubernetes.io/projected/d25b364f-96c6-4513-9142-c198d3a2cbae-kube-api-access-ltsz2\") pod \"d25b364f-96c6-4513-9142-c198d3a2cbae\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.574960 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-config\") pod \"d25b364f-96c6-4513-9142-c198d3a2cbae\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.575010 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d25b364f-96c6-4513-9142-c198d3a2cbae-serving-cert\") pod \"d25b364f-96c6-4513-9142-c198d3a2cbae\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.575034 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-client-ca\") pod \"d25b364f-96c6-4513-9142-c198d3a2cbae\" (UID: \"d25b364f-96c6-4513-9142-c198d3a2cbae\") " Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.575315 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.575332 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsrzh\" (UniqueName: \"kubernetes.io/projected/77a37156-be74-4f60-ad0d-9248408b80d4-kube-api-access-gsrzh\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.575342 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/77a37156-be74-4f60-ad0d-9248408b80d4-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.575372 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77a37156-be74-4f60-ad0d-9248408b80d4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.575590 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d25b364f-96c6-4513-9142-c198d3a2cbae" (UID: "d25b364f-96c6-4513-9142-c198d3a2cbae"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.575766 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-config" (OuterVolumeSpecName: "config") pod "d25b364f-96c6-4513-9142-c198d3a2cbae" (UID: "d25b364f-96c6-4513-9142-c198d3a2cbae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.575799 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-client-ca" (OuterVolumeSpecName: "client-ca") pod "d25b364f-96c6-4513-9142-c198d3a2cbae" (UID: "d25b364f-96c6-4513-9142-c198d3a2cbae"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.578511 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d25b364f-96c6-4513-9142-c198d3a2cbae-kube-api-access-ltsz2" (OuterVolumeSpecName: "kube-api-access-ltsz2") pod "d25b364f-96c6-4513-9142-c198d3a2cbae" (UID: "d25b364f-96c6-4513-9142-c198d3a2cbae"). InnerVolumeSpecName "kube-api-access-ltsz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.578714 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d25b364f-96c6-4513-9142-c198d3a2cbae-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d25b364f-96c6-4513-9142-c198d3a2cbae" (UID: "d25b364f-96c6-4513-9142-c198d3a2cbae"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.676818 4936 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.676860 4936 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.676877 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltsz2\" (UniqueName: \"kubernetes.io/projected/d25b364f-96c6-4513-9142-c198d3a2cbae-kube-api-access-ltsz2\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.676896 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d25b364f-96c6-4513-9142-c198d3a2cbae-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:53 crc kubenswrapper[4936]: I0320 16:05:53.676914 4936 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d25b364f-96c6-4513-9142-c198d3a2cbae-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.145347 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" event={"ID":"d25b364f-96c6-4513-9142-c198d3a2cbae","Type":"ContainerDied","Data":"7d99ba61184d77005a476969e3aa4eaaf6103c9f93bdfec53f96e4324099123d"} Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.145367 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8985447-scqrx" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.145404 4936 scope.go:117] "RemoveContainer" containerID="8728a72624f65e72cce1840b657ecc386792f304a89b0195276a6f7cc1ca7173" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.146625 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" event={"ID":"77a37156-be74-4f60-ad0d-9248408b80d4","Type":"ContainerDied","Data":"a6cf2ae5df35959c21646b48ac56ce371052b14984d6887cc6919dd004291fd4"} Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.146704 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.166616 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-scqrx"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.166909 4936 scope.go:117] "RemoveContainer" containerID="19ab13ddef3b485fe6601afb704083734d139b6c1aa2a27568617af553a0e75b" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.169293 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-scqrx"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.181112 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.203107 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7ff7586b44-mgrcq"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.487406 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95"] Mar 20 16:05:54 crc kubenswrapper[4936]: E0320 16:05:54.488040 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a37156-be74-4f60-ad0d-9248408b80d4" containerName="route-controller-manager" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.488062 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a37156-be74-4f60-ad0d-9248408b80d4" containerName="route-controller-manager" Mar 20 16:05:54 crc kubenswrapper[4936]: E0320 16:05:54.488077 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d25b364f-96c6-4513-9142-c198d3a2cbae" containerName="controller-manager" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.488087 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="d25b364f-96c6-4513-9142-c198d3a2cbae" containerName="controller-manager" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.488218 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="77a37156-be74-4f60-ad0d-9248408b80d4" containerName="route-controller-manager" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.488241 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="d25b364f-96c6-4513-9142-c198d3a2cbae" containerName="controller-manager" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.488797 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.492215 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74b956d8c5-tsp2c"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.492385 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.492390 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.492627 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.492638 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.492676 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.492630 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.492932 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.499350 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.499659 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.499691 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.499909 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.500016 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.500295 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.502369 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.504178 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.509612 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74b956d8c5-tsp2c"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.535584 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lq9hg"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.536640 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.544625 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.552955 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lq9hg"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.603177 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtw8w\" (UniqueName: \"kubernetes.io/projected/09dcc5da-7c70-4534-ab4c-265796bbc0d8-kube-api-access-qtw8w\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.603250 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbx7\" (UniqueName: \"kubernetes.io/projected/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-kube-api-access-vxbx7\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.603289 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-config\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.603316 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-proxy-ca-bundles\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.603341 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09dcc5da-7c70-4534-ab4c-265796bbc0d8-client-ca\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.603470 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09dcc5da-7c70-4534-ab4c-265796bbc0d8-serving-cert\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.603590 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-client-ca\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.603637 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09dcc5da-7c70-4534-ab4c-265796bbc0d8-config\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.603696 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-serving-cert\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.704715 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv7dz\" (UniqueName: \"kubernetes.io/projected/51fc6176-8334-4c16-bfd3-42584385ea47-kube-api-access-sv7dz\") pod \"redhat-operators-lq9hg\" (UID: \"51fc6176-8334-4c16-bfd3-42584385ea47\") " pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.705004 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-config\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.705107 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-proxy-ca-bundles\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.705198 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09dcc5da-7c70-4534-ab4c-265796bbc0d8-client-ca\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.705285 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fc6176-8334-4c16-bfd3-42584385ea47-utilities\") pod \"redhat-operators-lq9hg\" (UID: \"51fc6176-8334-4c16-bfd3-42584385ea47\") " pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.705405 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09dcc5da-7c70-4534-ab4c-265796bbc0d8-serving-cert\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.705558 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-client-ca\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.705685 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fc6176-8334-4c16-bfd3-42584385ea47-catalog-content\") pod \"redhat-operators-lq9hg\" (UID: \"51fc6176-8334-4c16-bfd3-42584385ea47\") " pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.705799 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09dcc5da-7c70-4534-ab4c-265796bbc0d8-config\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.705919 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-serving-cert\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.706054 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtw8w\" (UniqueName: \"kubernetes.io/projected/09dcc5da-7c70-4534-ab4c-265796bbc0d8-kube-api-access-qtw8w\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.706151 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbx7\" (UniqueName: \"kubernetes.io/projected/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-kube-api-access-vxbx7\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.706427 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09dcc5da-7c70-4534-ab4c-265796bbc0d8-client-ca\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.706518 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-client-ca\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.707138 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-proxy-ca-bundles\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.707558 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09dcc5da-7c70-4534-ab4c-265796bbc0d8-config\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.707632 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-config\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.710068 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-serving-cert\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.725523 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09dcc5da-7c70-4534-ab4c-265796bbc0d8-serving-cert\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.732099 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbx7\" (UniqueName: \"kubernetes.io/projected/e2b1b101-abcd-42c4-bb96-1cfaeacb8673-kube-api-access-vxbx7\") pod \"controller-manager-74b956d8c5-tsp2c\" (UID: \"e2b1b101-abcd-42c4-bb96-1cfaeacb8673\") " pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.737870 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h6c8f"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.741869 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.742472 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtw8w\" (UniqueName: \"kubernetes.io/projected/09dcc5da-7c70-4534-ab4c-265796bbc0d8-kube-api-access-qtw8w\") pod \"route-controller-manager-59577dcc5d-5hf95\" (UID: \"09dcc5da-7c70-4534-ab4c-265796bbc0d8\") " pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.743793 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.768365 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h6c8f"] Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.807885 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv7dz\" (UniqueName: \"kubernetes.io/projected/51fc6176-8334-4c16-bfd3-42584385ea47-kube-api-access-sv7dz\") pod \"redhat-operators-lq9hg\" (UID: \"51fc6176-8334-4c16-bfd3-42584385ea47\") " pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.807953 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fc6176-8334-4c16-bfd3-42584385ea47-utilities\") pod \"redhat-operators-lq9hg\" (UID: \"51fc6176-8334-4c16-bfd3-42584385ea47\") " pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.808007 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fc6176-8334-4c16-bfd3-42584385ea47-catalog-content\") pod \"redhat-operators-lq9hg\" (UID: \"51fc6176-8334-4c16-bfd3-42584385ea47\") " pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.808506 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fc6176-8334-4c16-bfd3-42584385ea47-catalog-content\") pod \"redhat-operators-lq9hg\" (UID: \"51fc6176-8334-4c16-bfd3-42584385ea47\") " pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.808750 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fc6176-8334-4c16-bfd3-42584385ea47-utilities\") pod \"redhat-operators-lq9hg\" (UID: \"51fc6176-8334-4c16-bfd3-42584385ea47\") " pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.809710 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.824852 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.828247 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv7dz\" (UniqueName: \"kubernetes.io/projected/51fc6176-8334-4c16-bfd3-42584385ea47-kube-api-access-sv7dz\") pod \"redhat-operators-lq9hg\" (UID: \"51fc6176-8334-4c16-bfd3-42584385ea47\") " pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.860167 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.908937 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1399ef65-8fa1-48a3-b7a2-c7ebd25198e3-utilities\") pod \"certified-operators-h6c8f\" (UID: \"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3\") " pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.908982 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrfsr\" (UniqueName: \"kubernetes.io/projected/1399ef65-8fa1-48a3-b7a2-c7ebd25198e3-kube-api-access-lrfsr\") pod \"certified-operators-h6c8f\" (UID: \"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3\") " pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:54 crc kubenswrapper[4936]: I0320 16:05:54.909004 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1399ef65-8fa1-48a3-b7a2-c7ebd25198e3-catalog-content\") pod \"certified-operators-h6c8f\" (UID: \"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3\") " pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.010603 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1399ef65-8fa1-48a3-b7a2-c7ebd25198e3-utilities\") pod \"certified-operators-h6c8f\" (UID: \"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3\") " pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.010655 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrfsr\" (UniqueName: \"kubernetes.io/projected/1399ef65-8fa1-48a3-b7a2-c7ebd25198e3-kube-api-access-lrfsr\") pod \"certified-operators-h6c8f\" (UID: \"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3\") " pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.010672 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1399ef65-8fa1-48a3-b7a2-c7ebd25198e3-catalog-content\") pod \"certified-operators-h6c8f\" (UID: \"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3\") " pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.011413 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1399ef65-8fa1-48a3-b7a2-c7ebd25198e3-catalog-content\") pod \"certified-operators-h6c8f\" (UID: \"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3\") " pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.012489 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1399ef65-8fa1-48a3-b7a2-c7ebd25198e3-utilities\") pod \"certified-operators-h6c8f\" (UID: \"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3\") " pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.030626 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrfsr\" (UniqueName: \"kubernetes.io/projected/1399ef65-8fa1-48a3-b7a2-c7ebd25198e3-kube-api-access-lrfsr\") pod \"certified-operators-h6c8f\" (UID: \"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3\") " pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.083275 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95"] Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.097164 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.213380 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" event={"ID":"09dcc5da-7c70-4534-ab4c-265796bbc0d8","Type":"ContainerStarted","Data":"ad2c2f1a3e45810e7b0dd67133285639eeeb976ff3e8123f1d0cbe1ce3a5da20"} Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.352770 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74b956d8c5-tsp2c"] Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.416988 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lq9hg"] Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.445131 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h6c8f"] Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.861944 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77a37156-be74-4f60-ad0d-9248408b80d4" path="/var/lib/kubelet/pods/77a37156-be74-4f60-ad0d-9248408b80d4/volumes" Mar 20 16:05:55 crc kubenswrapper[4936]: I0320 16:05:55.862871 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d25b364f-96c6-4513-9142-c198d3a2cbae" path="/var/lib/kubelet/pods/d25b364f-96c6-4513-9142-c198d3a2cbae/volumes" Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.227393 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" event={"ID":"e2b1b101-abcd-42c4-bb96-1cfaeacb8673","Type":"ContainerStarted","Data":"21e951a74cd1a6d49d4f3eeb24fb73994caae7031c0f43ca71cb37ffe8df9152"} Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.227756 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" event={"ID":"e2b1b101-abcd-42c4-bb96-1cfaeacb8673","Type":"ContainerStarted","Data":"a72c5ad25ff0407bffae033b1e94125a5a2f8ebf1430c22482478b1d9575ae7e"} Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.227922 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.230810 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" event={"ID":"09dcc5da-7c70-4534-ab4c-265796bbc0d8","Type":"ContainerStarted","Data":"cccd1a5db9acfa159ee64d6783a1a0ff7800aadfdc3784651d939b8823c160a8"} Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.231882 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.233137 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.234117 4936 generic.go:334] "Generic (PLEG): container finished" podID="1399ef65-8fa1-48a3-b7a2-c7ebd25198e3" containerID="f86d4a77686c8b3f827f9af0679fad38336a53cb9c457d7522da2f5bc8fb9eb4" exitCode=0 Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.234152 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6c8f" event={"ID":"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3","Type":"ContainerDied","Data":"f86d4a77686c8b3f827f9af0679fad38336a53cb9c457d7522da2f5bc8fb9eb4"} Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.234271 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6c8f" event={"ID":"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3","Type":"ContainerStarted","Data":"72770c31c28898a0dd7bfaa9bbe13f46812802b3d710ecbbc0afbb3bb39117f0"} Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.235620 4936 generic.go:334] "Generic (PLEG): container finished" podID="51fc6176-8334-4c16-bfd3-42584385ea47" containerID="953ba245fea2d5a09598aede0cef15083ae0d95004d0876d768f60fb311d3669" exitCode=0 Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.235655 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lq9hg" event={"ID":"51fc6176-8334-4c16-bfd3-42584385ea47","Type":"ContainerDied","Data":"953ba245fea2d5a09598aede0cef15083ae0d95004d0876d768f60fb311d3669"} Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.235673 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lq9hg" event={"ID":"51fc6176-8334-4c16-bfd3-42584385ea47","Type":"ContainerStarted","Data":"6ae43e2f19017df3df369b56283959014d19ce596f1c1e3b48da4c8a8b96c931"} Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.236783 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.255611 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74b956d8c5-tsp2c" podStartSLOduration=4.2555927 podStartE2EDuration="4.2555927s" podCreationTimestamp="2026-03-20 16:05:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:05:56.253983568 +0000 UTC m=+307.200351393" watchObservedRunningTime="2026-03-20 16:05:56.2555927 +0000 UTC m=+307.201960525" Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.345966 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59577dcc5d-5hf95" podStartSLOduration=4.345915405 podStartE2EDuration="4.345915405s" podCreationTimestamp="2026-03-20 16:05:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:05:56.340210924 +0000 UTC m=+307.286578749" watchObservedRunningTime="2026-03-20 16:05:56.345915405 +0000 UTC m=+307.292283220" Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.924721 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-84rkf"] Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.926617 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.929813 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 16:05:56 crc kubenswrapper[4936]: I0320 16:05:56.947581 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84rkf"] Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.046711 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/043c1af0-d920-45f8-b48c-fdc26c752e9a-catalog-content\") pod \"community-operators-84rkf\" (UID: \"043c1af0-d920-45f8-b48c-fdc26c752e9a\") " pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.046834 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r62x\" (UniqueName: \"kubernetes.io/projected/043c1af0-d920-45f8-b48c-fdc26c752e9a-kube-api-access-2r62x\") pod \"community-operators-84rkf\" (UID: \"043c1af0-d920-45f8-b48c-fdc26c752e9a\") " pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.046908 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/043c1af0-d920-45f8-b48c-fdc26c752e9a-utilities\") pod \"community-operators-84rkf\" (UID: \"043c1af0-d920-45f8-b48c-fdc26c752e9a\") " pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.136458 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tncvv"] Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.138703 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.143832 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tncvv"] Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.144312 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.149735 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/043c1af0-d920-45f8-b48c-fdc26c752e9a-utilities\") pod \"community-operators-84rkf\" (UID: \"043c1af0-d920-45f8-b48c-fdc26c752e9a\") " pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.149815 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/043c1af0-d920-45f8-b48c-fdc26c752e9a-catalog-content\") pod \"community-operators-84rkf\" (UID: \"043c1af0-d920-45f8-b48c-fdc26c752e9a\") " pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.149846 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r62x\" (UniqueName: \"kubernetes.io/projected/043c1af0-d920-45f8-b48c-fdc26c752e9a-kube-api-access-2r62x\") pod \"community-operators-84rkf\" (UID: \"043c1af0-d920-45f8-b48c-fdc26c752e9a\") " pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.151114 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/043c1af0-d920-45f8-b48c-fdc26c752e9a-utilities\") pod \"community-operators-84rkf\" (UID: \"043c1af0-d920-45f8-b48c-fdc26c752e9a\") " pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.152416 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/043c1af0-d920-45f8-b48c-fdc26c752e9a-catalog-content\") pod \"community-operators-84rkf\" (UID: \"043c1af0-d920-45f8-b48c-fdc26c752e9a\") " pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.172236 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r62x\" (UniqueName: \"kubernetes.io/projected/043c1af0-d920-45f8-b48c-fdc26c752e9a-kube-api-access-2r62x\") pod \"community-operators-84rkf\" (UID: \"043c1af0-d920-45f8-b48c-fdc26c752e9a\") " pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.251184 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0083ee-8688-481a-86c6-df9338ea1364-catalog-content\") pod \"redhat-marketplace-tncvv\" (UID: \"bb0083ee-8688-481a-86c6-df9338ea1364\") " pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.251302 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j25mx\" (UniqueName: \"kubernetes.io/projected/bb0083ee-8688-481a-86c6-df9338ea1364-kube-api-access-j25mx\") pod \"redhat-marketplace-tncvv\" (UID: \"bb0083ee-8688-481a-86c6-df9338ea1364\") " pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.251336 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0083ee-8688-481a-86c6-df9338ea1364-utilities\") pod \"redhat-marketplace-tncvv\" (UID: \"bb0083ee-8688-481a-86c6-df9338ea1364\") " pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.353048 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0083ee-8688-481a-86c6-df9338ea1364-catalog-content\") pod \"redhat-marketplace-tncvv\" (UID: \"bb0083ee-8688-481a-86c6-df9338ea1364\") " pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.353365 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j25mx\" (UniqueName: \"kubernetes.io/projected/bb0083ee-8688-481a-86c6-df9338ea1364-kube-api-access-j25mx\") pod \"redhat-marketplace-tncvv\" (UID: \"bb0083ee-8688-481a-86c6-df9338ea1364\") " pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.353399 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0083ee-8688-481a-86c6-df9338ea1364-utilities\") pod \"redhat-marketplace-tncvv\" (UID: \"bb0083ee-8688-481a-86c6-df9338ea1364\") " pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.353880 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0083ee-8688-481a-86c6-df9338ea1364-catalog-content\") pod \"redhat-marketplace-tncvv\" (UID: \"bb0083ee-8688-481a-86c6-df9338ea1364\") " pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.354441 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0083ee-8688-481a-86c6-df9338ea1364-utilities\") pod \"redhat-marketplace-tncvv\" (UID: \"bb0083ee-8688-481a-86c6-df9338ea1364\") " pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.373888 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j25mx\" (UniqueName: \"kubernetes.io/projected/bb0083ee-8688-481a-86c6-df9338ea1364-kube-api-access-j25mx\") pod \"redhat-marketplace-tncvv\" (UID: \"bb0083ee-8688-481a-86c6-df9338ea1364\") " pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.608674 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:05:57 crc kubenswrapper[4936]: I0320 16:05:57.663718 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.060942 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84rkf"] Mar 20 16:05:58 crc kubenswrapper[4936]: W0320 16:05:58.064899 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod043c1af0_d920_45f8_b48c_fdc26c752e9a.slice/crio-8477945e60ae2fff92b7cd4ed85a387de484856d9857e8a2117d12042ac61d8d WatchSource:0}: Error finding container 8477945e60ae2fff92b7cd4ed85a387de484856d9857e8a2117d12042ac61d8d: Status 404 returned error can't find the container with id 8477945e60ae2fff92b7cd4ed85a387de484856d9857e8a2117d12042ac61d8d Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.148155 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tncvv"] Mar 20 16:05:58 crc kubenswrapper[4936]: W0320 16:05:58.154720 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb0083ee_8688_481a_86c6_df9338ea1364.slice/crio-01b82a4e9eeb5d5b461a8ce7154e54beddc7b119e016b8bf334d05af3c88d858 WatchSource:0}: Error finding container 01b82a4e9eeb5d5b461a8ce7154e54beddc7b119e016b8bf334d05af3c88d858: Status 404 returned error can't find the container with id 01b82a4e9eeb5d5b461a8ce7154e54beddc7b119e016b8bf334d05af3c88d858 Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.251343 4936 generic.go:334] "Generic (PLEG): container finished" podID="1399ef65-8fa1-48a3-b7a2-c7ebd25198e3" containerID="eea6b3707c8268b79ded01327a199f4558ff21de938b14567208d0e1b2f178a9" exitCode=0 Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.251784 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6c8f" event={"ID":"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3","Type":"ContainerDied","Data":"eea6b3707c8268b79ded01327a199f4558ff21de938b14567208d0e1b2f178a9"} Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.264922 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lq9hg" event={"ID":"51fc6176-8334-4c16-bfd3-42584385ea47","Type":"ContainerStarted","Data":"4ba708a2d0f6b60b4f50a46f8d6d73869cb638c64acb8b9ea82103c31e200e2d"} Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.268050 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tncvv" event={"ID":"bb0083ee-8688-481a-86c6-df9338ea1364","Type":"ContainerStarted","Data":"01b82a4e9eeb5d5b461a8ce7154e54beddc7b119e016b8bf334d05af3c88d858"} Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.270623 4936 generic.go:334] "Generic (PLEG): container finished" podID="043c1af0-d920-45f8-b48c-fdc26c752e9a" containerID="ee625ca8f3be5ecc24baf86f2d749a7e3f2a5a6e65c9b13c736e9131c891312c" exitCode=0 Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.270825 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84rkf" event={"ID":"043c1af0-d920-45f8-b48c-fdc26c752e9a","Type":"ContainerDied","Data":"ee625ca8f3be5ecc24baf86f2d749a7e3f2a5a6e65c9b13c736e9131c891312c"} Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.270890 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84rkf" event={"ID":"043c1af0-d920-45f8-b48c-fdc26c752e9a","Type":"ContainerStarted","Data":"8477945e60ae2fff92b7cd4ed85a387de484856d9857e8a2117d12042ac61d8d"} Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.854660 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.854714 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.854766 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.855429 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"548766dd512244399f51b55b943c4d13acad4bf13ce80ec19f3b99a712997270"} pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 16:05:58 crc kubenswrapper[4936]: I0320 16:05:58.855501 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" containerID="cri-o://548766dd512244399f51b55b943c4d13acad4bf13ce80ec19f3b99a712997270" gracePeriod=600 Mar 20 16:05:59 crc kubenswrapper[4936]: I0320 16:05:59.278170 4936 generic.go:334] "Generic (PLEG): container finished" podID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerID="548766dd512244399f51b55b943c4d13acad4bf13ce80ec19f3b99a712997270" exitCode=0 Mar 20 16:05:59 crc kubenswrapper[4936]: I0320 16:05:59.278250 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerDied","Data":"548766dd512244399f51b55b943c4d13acad4bf13ce80ec19f3b99a712997270"} Mar 20 16:05:59 crc kubenswrapper[4936]: I0320 16:05:59.285471 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6c8f" event={"ID":"1399ef65-8fa1-48a3-b7a2-c7ebd25198e3","Type":"ContainerStarted","Data":"ebb5a300fefc9057a62f8ad9cf9ff5815c709264b1e7551123794b92fc04893f"} Mar 20 16:05:59 crc kubenswrapper[4936]: I0320 16:05:59.287356 4936 generic.go:334] "Generic (PLEG): container finished" podID="51fc6176-8334-4c16-bfd3-42584385ea47" containerID="4ba708a2d0f6b60b4f50a46f8d6d73869cb638c64acb8b9ea82103c31e200e2d" exitCode=0 Mar 20 16:05:59 crc kubenswrapper[4936]: I0320 16:05:59.287416 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lq9hg" event={"ID":"51fc6176-8334-4c16-bfd3-42584385ea47","Type":"ContainerDied","Data":"4ba708a2d0f6b60b4f50a46f8d6d73869cb638c64acb8b9ea82103c31e200e2d"} Mar 20 16:05:59 crc kubenswrapper[4936]: I0320 16:05:59.289047 4936 generic.go:334] "Generic (PLEG): container finished" podID="bb0083ee-8688-481a-86c6-df9338ea1364" containerID="225ea47ad16447520d09acaba428a2eb82ce7218bd27e88ceca828c1554072dc" exitCode=0 Mar 20 16:05:59 crc kubenswrapper[4936]: I0320 16:05:59.289096 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tncvv" event={"ID":"bb0083ee-8688-481a-86c6-df9338ea1364","Type":"ContainerDied","Data":"225ea47ad16447520d09acaba428a2eb82ce7218bd27e88ceca828c1554072dc"} Mar 20 16:05:59 crc kubenswrapper[4936]: I0320 16:05:59.307660 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h6c8f" podStartSLOduration=2.850960783 podStartE2EDuration="5.307636039s" podCreationTimestamp="2026-03-20 16:05:54 +0000 UTC" firstStartedPulling="2026-03-20 16:05:56.237634754 +0000 UTC m=+307.184002569" lastFinishedPulling="2026-03-20 16:05:58.69431001 +0000 UTC m=+309.640677825" observedRunningTime="2026-03-20 16:05:59.304445934 +0000 UTC m=+310.250813759" watchObservedRunningTime="2026-03-20 16:05:59.307636039 +0000 UTC m=+310.254003854" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.161620 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567046-nr4fn"] Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.163073 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567046-nr4fn" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.164744 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.165118 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.167665 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.175149 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567046-nr4fn"] Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.298000 4936 generic.go:334] "Generic (PLEG): container finished" podID="043c1af0-d920-45f8-b48c-fdc26c752e9a" containerID="c744e876b858da9f76bed066828187d2e755029f96154771aee6c72e44c5b7aa" exitCode=0 Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.298105 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84rkf" event={"ID":"043c1af0-d920-45f8-b48c-fdc26c752e9a","Type":"ContainerDied","Data":"c744e876b858da9f76bed066828187d2e755029f96154771aee6c72e44c5b7aa"} Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.301684 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"ad7190bb15359c2cb63ff782fee447cfc84994fffc54cec84eef49413cfae9d9"} Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.308139 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lq9hg" event={"ID":"51fc6176-8334-4c16-bfd3-42584385ea47","Type":"ContainerStarted","Data":"8aff595bfc982553bd2647547671ce81dcdc54fbe67c432db9a1e880271faf88"} Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.358810 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7c64\" (UniqueName: \"kubernetes.io/projected/a1ffe995-5ea7-46df-9389-1ac1e6bd729b-kube-api-access-q7c64\") pod \"auto-csr-approver-29567046-nr4fn\" (UID: \"a1ffe995-5ea7-46df-9389-1ac1e6bd729b\") " pod="openshift-infra/auto-csr-approver-29567046-nr4fn" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.392314 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lq9hg" podStartSLOduration=2.905766146 podStartE2EDuration="6.392275993s" podCreationTimestamp="2026-03-20 16:05:54 +0000 UTC" firstStartedPulling="2026-03-20 16:05:56.237670025 +0000 UTC m=+307.184037840" lastFinishedPulling="2026-03-20 16:05:59.724179862 +0000 UTC m=+310.670547687" observedRunningTime="2026-03-20 16:06:00.383966162 +0000 UTC m=+311.330333977" watchObservedRunningTime="2026-03-20 16:06:00.392275993 +0000 UTC m=+311.338643808" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.461180 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7c64\" (UniqueName: \"kubernetes.io/projected/a1ffe995-5ea7-46df-9389-1ac1e6bd729b-kube-api-access-q7c64\") pod \"auto-csr-approver-29567046-nr4fn\" (UID: \"a1ffe995-5ea7-46df-9389-1ac1e6bd729b\") " pod="openshift-infra/auto-csr-approver-29567046-nr4fn" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.492442 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7c64\" (UniqueName: \"kubernetes.io/projected/a1ffe995-5ea7-46df-9389-1ac1e6bd729b-kube-api-access-q7c64\") pod \"auto-csr-approver-29567046-nr4fn\" (UID: \"a1ffe995-5ea7-46df-9389-1ac1e6bd729b\") " pod="openshift-infra/auto-csr-approver-29567046-nr4fn" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.518438 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567046-nr4fn" Mar 20 16:06:00 crc kubenswrapper[4936]: I0320 16:06:00.960880 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567046-nr4fn"] Mar 20 16:06:01 crc kubenswrapper[4936]: I0320 16:06:01.317642 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84rkf" event={"ID":"043c1af0-d920-45f8-b48c-fdc26c752e9a","Type":"ContainerStarted","Data":"4392a92269ffd85433a470a5eb39a6bc659036d65871e501d1c8ee98f5503236"} Mar 20 16:06:01 crc kubenswrapper[4936]: I0320 16:06:01.321914 4936 generic.go:334] "Generic (PLEG): container finished" podID="bb0083ee-8688-481a-86c6-df9338ea1364" containerID="472b0274550d73719ab356d979525b293077bd7286e06cc31fbf82976d8029a0" exitCode=0 Mar 20 16:06:01 crc kubenswrapper[4936]: I0320 16:06:01.321976 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tncvv" event={"ID":"bb0083ee-8688-481a-86c6-df9338ea1364","Type":"ContainerDied","Data":"472b0274550d73719ab356d979525b293077bd7286e06cc31fbf82976d8029a0"} Mar 20 16:06:01 crc kubenswrapper[4936]: I0320 16:06:01.323638 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567046-nr4fn" event={"ID":"a1ffe995-5ea7-46df-9389-1ac1e6bd729b","Type":"ContainerStarted","Data":"853b3591d6ec658df4ee9691023729bd93534bf3a00658d48bc1fa955fdfd902"} Mar 20 16:06:01 crc kubenswrapper[4936]: I0320 16:06:01.349059 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-84rkf" podStartSLOduration=2.822317573 podStartE2EDuration="5.349041676s" podCreationTimestamp="2026-03-20 16:05:56 +0000 UTC" firstStartedPulling="2026-03-20 16:05:58.277135831 +0000 UTC m=+309.223503646" lastFinishedPulling="2026-03-20 16:06:00.803859934 +0000 UTC m=+311.750227749" observedRunningTime="2026-03-20 16:06:01.34692514 +0000 UTC m=+312.293292995" watchObservedRunningTime="2026-03-20 16:06:01.349041676 +0000 UTC m=+312.295409491" Mar 20 16:06:02 crc kubenswrapper[4936]: I0320 16:06:02.342337 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tncvv" event={"ID":"bb0083ee-8688-481a-86c6-df9338ea1364","Type":"ContainerStarted","Data":"db33f37ebf090ed7f79179922d45fe472e1154d4e945e3830f95a74466d0fbf7"} Mar 20 16:06:02 crc kubenswrapper[4936]: I0320 16:06:02.344533 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567046-nr4fn" event={"ID":"a1ffe995-5ea7-46df-9389-1ac1e6bd729b","Type":"ContainerStarted","Data":"e7773d332dfd7e5acd15f8dc7a3466a64370e0e72622fea11316295c4202d5d2"} Mar 20 16:06:02 crc kubenswrapper[4936]: I0320 16:06:02.361357 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tncvv" podStartSLOduration=2.917588609 podStartE2EDuration="5.361331561s" podCreationTimestamp="2026-03-20 16:05:57 +0000 UTC" firstStartedPulling="2026-03-20 16:05:59.290660709 +0000 UTC m=+310.237028534" lastFinishedPulling="2026-03-20 16:06:01.734403671 +0000 UTC m=+312.680771486" observedRunningTime="2026-03-20 16:06:02.35941084 +0000 UTC m=+313.305778675" watchObservedRunningTime="2026-03-20 16:06:02.361331561 +0000 UTC m=+313.307699396" Mar 20 16:06:02 crc kubenswrapper[4936]: I0320 16:06:02.383792 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567046-nr4fn" podStartSLOduration=1.447940928 podStartE2EDuration="2.383775706s" podCreationTimestamp="2026-03-20 16:06:00 +0000 UTC" firstStartedPulling="2026-03-20 16:06:00.99234514 +0000 UTC m=+311.938712945" lastFinishedPulling="2026-03-20 16:06:01.928179918 +0000 UTC m=+312.874547723" observedRunningTime="2026-03-20 16:06:02.375107706 +0000 UTC m=+313.321475521" watchObservedRunningTime="2026-03-20 16:06:02.383775706 +0000 UTC m=+313.330143521" Mar 20 16:06:03 crc kubenswrapper[4936]: I0320 16:06:03.352415 4936 generic.go:334] "Generic (PLEG): container finished" podID="a1ffe995-5ea7-46df-9389-1ac1e6bd729b" containerID="e7773d332dfd7e5acd15f8dc7a3466a64370e0e72622fea11316295c4202d5d2" exitCode=0 Mar 20 16:06:03 crc kubenswrapper[4936]: I0320 16:06:03.352525 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567046-nr4fn" event={"ID":"a1ffe995-5ea7-46df-9389-1ac1e6bd729b","Type":"ContainerDied","Data":"e7773d332dfd7e5acd15f8dc7a3466a64370e0e72622fea11316295c4202d5d2"} Mar 20 16:06:04 crc kubenswrapper[4936]: I0320 16:06:04.766421 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567046-nr4fn" Mar 20 16:06:04 crc kubenswrapper[4936]: I0320 16:06:04.860943 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:06:04 crc kubenswrapper[4936]: I0320 16:06:04.861004 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:06:04 crc kubenswrapper[4936]: I0320 16:06:04.934387 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7c64\" (UniqueName: \"kubernetes.io/projected/a1ffe995-5ea7-46df-9389-1ac1e6bd729b-kube-api-access-q7c64\") pod \"a1ffe995-5ea7-46df-9389-1ac1e6bd729b\" (UID: \"a1ffe995-5ea7-46df-9389-1ac1e6bd729b\") " Mar 20 16:06:04 crc kubenswrapper[4936]: I0320 16:06:04.941101 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ffe995-5ea7-46df-9389-1ac1e6bd729b-kube-api-access-q7c64" (OuterVolumeSpecName: "kube-api-access-q7c64") pod "a1ffe995-5ea7-46df-9389-1ac1e6bd729b" (UID: "a1ffe995-5ea7-46df-9389-1ac1e6bd729b"). InnerVolumeSpecName "kube-api-access-q7c64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:06:05 crc kubenswrapper[4936]: I0320 16:06:05.036512 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7c64\" (UniqueName: \"kubernetes.io/projected/a1ffe995-5ea7-46df-9389-1ac1e6bd729b-kube-api-access-q7c64\") on node \"crc\" DevicePath \"\"" Mar 20 16:06:05 crc kubenswrapper[4936]: I0320 16:06:05.098859 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:06:05 crc kubenswrapper[4936]: I0320 16:06:05.098934 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:06:05 crc kubenswrapper[4936]: I0320 16:06:05.163560 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:06:05 crc kubenswrapper[4936]: I0320 16:06:05.367011 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567046-nr4fn" event={"ID":"a1ffe995-5ea7-46df-9389-1ac1e6bd729b","Type":"ContainerDied","Data":"853b3591d6ec658df4ee9691023729bd93534bf3a00658d48bc1fa955fdfd902"} Mar 20 16:06:05 crc kubenswrapper[4936]: I0320 16:06:05.367065 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="853b3591d6ec658df4ee9691023729bd93534bf3a00658d48bc1fa955fdfd902" Mar 20 16:06:05 crc kubenswrapper[4936]: I0320 16:06:05.367039 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567046-nr4fn" Mar 20 16:06:05 crc kubenswrapper[4936]: I0320 16:06:05.415512 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h6c8f" Mar 20 16:06:05 crc kubenswrapper[4936]: I0320 16:06:05.910775 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lq9hg" podUID="51fc6176-8334-4c16-bfd3-42584385ea47" containerName="registry-server" probeResult="failure" output=< Mar 20 16:06:05 crc kubenswrapper[4936]: timeout: failed to connect service ":50051" within 1s Mar 20 16:06:05 crc kubenswrapper[4936]: > Mar 20 16:06:07 crc kubenswrapper[4936]: I0320 16:06:07.609980 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:06:07 crc kubenswrapper[4936]: I0320 16:06:07.610325 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:06:07 crc kubenswrapper[4936]: I0320 16:06:07.662251 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:06:07 crc kubenswrapper[4936]: I0320 16:06:07.665826 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:06:07 crc kubenswrapper[4936]: I0320 16:06:07.665877 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:06:07 crc kubenswrapper[4936]: I0320 16:06:07.724782 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:06:08 crc kubenswrapper[4936]: I0320 16:06:08.430971 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tncvv" Mar 20 16:06:08 crc kubenswrapper[4936]: I0320 16:06:08.431736 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-84rkf" Mar 20 16:06:14 crc kubenswrapper[4936]: I0320 16:06:14.907349 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:06:14 crc kubenswrapper[4936]: I0320 16:06:14.951431 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lq9hg" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.701732 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dwqr9"] Mar 20 16:06:17 crc kubenswrapper[4936]: E0320 16:06:17.702232 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ffe995-5ea7-46df-9389-1ac1e6bd729b" containerName="oc" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.702247 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ffe995-5ea7-46df-9389-1ac1e6bd729b" containerName="oc" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.702365 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ffe995-5ea7-46df-9389-1ac1e6bd729b" containerName="oc" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.702900 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.723501 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dwqr9"] Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.838256 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.838302 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9d3625e4-1df5-4d64-ab02-238b89611894-registry-certificates\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.838327 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9d3625e4-1df5-4d64-ab02-238b89611894-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.838347 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9d3625e4-1df5-4d64-ab02-238b89611894-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.838385 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9d3625e4-1df5-4d64-ab02-238b89611894-registry-tls\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.838420 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9d3625e4-1df5-4d64-ab02-238b89611894-bound-sa-token\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.838445 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d3625e4-1df5-4d64-ab02-238b89611894-trusted-ca\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.838461 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqqtg\" (UniqueName: \"kubernetes.io/projected/9d3625e4-1df5-4d64-ab02-238b89611894-kube-api-access-xqqtg\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.859663 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.940175 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9d3625e4-1df5-4d64-ab02-238b89611894-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.940244 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9d3625e4-1df5-4d64-ab02-238b89611894-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.940323 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9d3625e4-1df5-4d64-ab02-238b89611894-registry-tls\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.940397 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9d3625e4-1df5-4d64-ab02-238b89611894-bound-sa-token\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.940431 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d3625e4-1df5-4d64-ab02-238b89611894-trusted-ca\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.940469 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqqtg\" (UniqueName: \"kubernetes.io/projected/9d3625e4-1df5-4d64-ab02-238b89611894-kube-api-access-xqqtg\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.940575 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9d3625e4-1df5-4d64-ab02-238b89611894-registry-certificates\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.941592 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9d3625e4-1df5-4d64-ab02-238b89611894-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.943519 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9d3625e4-1df5-4d64-ab02-238b89611894-registry-certificates\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.944818 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d3625e4-1df5-4d64-ab02-238b89611894-trusted-ca\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.947051 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9d3625e4-1df5-4d64-ab02-238b89611894-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.947620 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9d3625e4-1df5-4d64-ab02-238b89611894-registry-tls\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.965777 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9d3625e4-1df5-4d64-ab02-238b89611894-bound-sa-token\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:17 crc kubenswrapper[4936]: I0320 16:06:17.966726 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqqtg\" (UniqueName: \"kubernetes.io/projected/9d3625e4-1df5-4d64-ab02-238b89611894-kube-api-access-xqqtg\") pod \"image-registry-66df7c8f76-dwqr9\" (UID: \"9d3625e4-1df5-4d64-ab02-238b89611894\") " pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:18 crc kubenswrapper[4936]: I0320 16:06:18.024225 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:18 crc kubenswrapper[4936]: I0320 16:06:18.472284 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dwqr9"] Mar 20 16:06:18 crc kubenswrapper[4936]: W0320 16:06:18.484118 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d3625e4_1df5_4d64_ab02_238b89611894.slice/crio-7bbfbfdebb20dac5c767b1fffed4ad3b37518e94017b23bbacf2a09f3f42974a WatchSource:0}: Error finding container 7bbfbfdebb20dac5c767b1fffed4ad3b37518e94017b23bbacf2a09f3f42974a: Status 404 returned error can't find the container with id 7bbfbfdebb20dac5c767b1fffed4ad3b37518e94017b23bbacf2a09f3f42974a Mar 20 16:06:19 crc kubenswrapper[4936]: I0320 16:06:19.461772 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" event={"ID":"9d3625e4-1df5-4d64-ab02-238b89611894","Type":"ContainerStarted","Data":"f2b8f6b90a7200d2f3cfdae316ecedadc9991f9f03774fe036ea67a394a8ddc2"} Mar 20 16:06:19 crc kubenswrapper[4936]: I0320 16:06:19.462106 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" event={"ID":"9d3625e4-1df5-4d64-ab02-238b89611894","Type":"ContainerStarted","Data":"7bbfbfdebb20dac5c767b1fffed4ad3b37518e94017b23bbacf2a09f3f42974a"} Mar 20 16:06:19 crc kubenswrapper[4936]: I0320 16:06:19.463068 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:19 crc kubenswrapper[4936]: I0320 16:06:19.488451 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" podStartSLOduration=2.488431353 podStartE2EDuration="2.488431353s" podCreationTimestamp="2026-03-20 16:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:06:19.484096299 +0000 UTC m=+330.430464134" watchObservedRunningTime="2026-03-20 16:06:19.488431353 +0000 UTC m=+330.434799168" Mar 20 16:06:38 crc kubenswrapper[4936]: I0320 16:06:38.032336 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-dwqr9" Mar 20 16:06:38 crc kubenswrapper[4936]: I0320 16:06:38.111860 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8tlt4"] Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.175149 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" podUID="53d1ed6e-605c-46b9-b053-ba5a899db885" containerName="registry" containerID="cri-o://ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14" gracePeriod=30 Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.317402 4936 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-8tlt4 container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.19:5000/healthz\": dial tcp 10.217.0.19:5000: connect: connection refused" start-of-body= Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.317494 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" podUID="53d1ed6e-605c-46b9-b053-ba5a899db885" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.19:5000/healthz\": dial tcp 10.217.0.19:5000: connect: connection refused" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.641622 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.714836 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"53d1ed6e-605c-46b9-b053-ba5a899db885\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.714932 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g45pg\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-kube-api-access-g45pg\") pod \"53d1ed6e-605c-46b9-b053-ba5a899db885\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.714975 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-tls\") pod \"53d1ed6e-605c-46b9-b053-ba5a899db885\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.715004 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53d1ed6e-605c-46b9-b053-ba5a899db885-ca-trust-extracted\") pod \"53d1ed6e-605c-46b9-b053-ba5a899db885\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.715021 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53d1ed6e-605c-46b9-b053-ba5a899db885-installation-pull-secrets\") pod \"53d1ed6e-605c-46b9-b053-ba5a899db885\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.715036 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-certificates\") pod \"53d1ed6e-605c-46b9-b053-ba5a899db885\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.715052 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-bound-sa-token\") pod \"53d1ed6e-605c-46b9-b053-ba5a899db885\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.715117 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-trusted-ca\") pod \"53d1ed6e-605c-46b9-b053-ba5a899db885\" (UID: \"53d1ed6e-605c-46b9-b053-ba5a899db885\") " Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.716289 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "53d1ed6e-605c-46b9-b053-ba5a899db885" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.716729 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "53d1ed6e-605c-46b9-b053-ba5a899db885" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.721941 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "53d1ed6e-605c-46b9-b053-ba5a899db885" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.721973 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d1ed6e-605c-46b9-b053-ba5a899db885-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "53d1ed6e-605c-46b9-b053-ba5a899db885" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.722276 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "53d1ed6e-605c-46b9-b053-ba5a899db885" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.722878 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-kube-api-access-g45pg" (OuterVolumeSpecName: "kube-api-access-g45pg") pod "53d1ed6e-605c-46b9-b053-ba5a899db885" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885"). InnerVolumeSpecName "kube-api-access-g45pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.728836 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "53d1ed6e-605c-46b9-b053-ba5a899db885" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.730853 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53d1ed6e-605c-46b9-b053-ba5a899db885-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "53d1ed6e-605c-46b9-b053-ba5a899db885" (UID: "53d1ed6e-605c-46b9-b053-ba5a899db885"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.788663 4936 generic.go:334] "Generic (PLEG): container finished" podID="53d1ed6e-605c-46b9-b053-ba5a899db885" containerID="ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14" exitCode=0 Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.788722 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.788724 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" event={"ID":"53d1ed6e-605c-46b9-b053-ba5a899db885","Type":"ContainerDied","Data":"ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14"} Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.788775 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8tlt4" event={"ID":"53d1ed6e-605c-46b9-b053-ba5a899db885","Type":"ContainerDied","Data":"39e74f84a3982bf3cb8fadb8bc8e4cdf799df704fd09e84c35e6a30448ba9d57"} Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.788864 4936 scope.go:117] "RemoveContainer" containerID="ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.816611 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g45pg\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-kube-api-access-g45pg\") on node \"crc\" DevicePath \"\"" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.816649 4936 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.816665 4936 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53d1ed6e-605c-46b9-b053-ba5a899db885-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.816677 4936 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53d1ed6e-605c-46b9-b053-ba5a899db885-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.816689 4936 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.816700 4936 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53d1ed6e-605c-46b9-b053-ba5a899db885-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.816710 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53d1ed6e-605c-46b9-b053-ba5a899db885-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.824190 4936 scope.go:117] "RemoveContainer" containerID="ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14" Mar 20 16:07:03 crc kubenswrapper[4936]: E0320 16:07:03.831462 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14\": container with ID starting with ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14 not found: ID does not exist" containerID="ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.831527 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14"} err="failed to get container status \"ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14\": rpc error: code = NotFound desc = could not find container \"ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14\": container with ID starting with ba1ab7c7201c434bb9a35b958388bbf911d29be421596ce3068ffbcd18f0bd14 not found: ID does not exist" Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.840619 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8tlt4"] Mar 20 16:07:03 crc kubenswrapper[4936]: I0320 16:07:03.859988 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8tlt4"] Mar 20 16:07:05 crc kubenswrapper[4936]: I0320 16:07:05.866441 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d1ed6e-605c-46b9-b053-ba5a899db885" path="/var/lib/kubelet/pods/53d1ed6e-605c-46b9-b053-ba5a899db885/volumes" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.142682 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567048-c4c4x"] Mar 20 16:08:00 crc kubenswrapper[4936]: E0320 16:08:00.145822 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d1ed6e-605c-46b9-b053-ba5a899db885" containerName="registry" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.145846 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d1ed6e-605c-46b9-b053-ba5a899db885" containerName="registry" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.145941 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d1ed6e-605c-46b9-b053-ba5a899db885" containerName="registry" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.146484 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567048-c4c4x" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.149721 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.150643 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.151619 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.155291 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567048-c4c4x"] Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.284295 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5x65\" (UniqueName: \"kubernetes.io/projected/e10e6f00-e11c-4f35-abad-7127db0c7cf3-kube-api-access-w5x65\") pod \"auto-csr-approver-29567048-c4c4x\" (UID: \"e10e6f00-e11c-4f35-abad-7127db0c7cf3\") " pod="openshift-infra/auto-csr-approver-29567048-c4c4x" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.385276 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5x65\" (UniqueName: \"kubernetes.io/projected/e10e6f00-e11c-4f35-abad-7127db0c7cf3-kube-api-access-w5x65\") pod \"auto-csr-approver-29567048-c4c4x\" (UID: \"e10e6f00-e11c-4f35-abad-7127db0c7cf3\") " pod="openshift-infra/auto-csr-approver-29567048-c4c4x" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.414320 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5x65\" (UniqueName: \"kubernetes.io/projected/e10e6f00-e11c-4f35-abad-7127db0c7cf3-kube-api-access-w5x65\") pod \"auto-csr-approver-29567048-c4c4x\" (UID: \"e10e6f00-e11c-4f35-abad-7127db0c7cf3\") " pod="openshift-infra/auto-csr-approver-29567048-c4c4x" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.472217 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567048-c4c4x" Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.726969 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567048-c4c4x"] Mar 20 16:08:00 crc kubenswrapper[4936]: W0320 16:08:00.749685 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode10e6f00_e11c_4f35_abad_7127db0c7cf3.slice/crio-e42a3bd0662758e76439d2f4ea698f43efa9368fbd6551c77c7e48336e5baf97 WatchSource:0}: Error finding container e42a3bd0662758e76439d2f4ea698f43efa9368fbd6551c77c7e48336e5baf97: Status 404 returned error can't find the container with id e42a3bd0662758e76439d2f4ea698f43efa9368fbd6551c77c7e48336e5baf97 Mar 20 16:08:00 crc kubenswrapper[4936]: I0320 16:08:00.756299 4936 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 16:08:01 crc kubenswrapper[4936]: I0320 16:08:01.199254 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567048-c4c4x" event={"ID":"e10e6f00-e11c-4f35-abad-7127db0c7cf3","Type":"ContainerStarted","Data":"e42a3bd0662758e76439d2f4ea698f43efa9368fbd6551c77c7e48336e5baf97"} Mar 20 16:08:02 crc kubenswrapper[4936]: I0320 16:08:02.207980 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567048-c4c4x" event={"ID":"e10e6f00-e11c-4f35-abad-7127db0c7cf3","Type":"ContainerStarted","Data":"e5f3de3dfb82abd886df377c901d821179552144b341c3b1fa6b7f45bc7753e6"} Mar 20 16:08:02 crc kubenswrapper[4936]: I0320 16:08:02.225080 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567048-c4c4x" podStartSLOduration=1.17568867 podStartE2EDuration="2.225026046s" podCreationTimestamp="2026-03-20 16:08:00 +0000 UTC" firstStartedPulling="2026-03-20 16:08:00.75584437 +0000 UTC m=+431.702212225" lastFinishedPulling="2026-03-20 16:08:01.805181746 +0000 UTC m=+432.751549601" observedRunningTime="2026-03-20 16:08:02.223704286 +0000 UTC m=+433.170072101" watchObservedRunningTime="2026-03-20 16:08:02.225026046 +0000 UTC m=+433.171393901" Mar 20 16:08:03 crc kubenswrapper[4936]: I0320 16:08:03.217948 4936 generic.go:334] "Generic (PLEG): container finished" podID="e10e6f00-e11c-4f35-abad-7127db0c7cf3" containerID="e5f3de3dfb82abd886df377c901d821179552144b341c3b1fa6b7f45bc7753e6" exitCode=0 Mar 20 16:08:03 crc kubenswrapper[4936]: I0320 16:08:03.218028 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567048-c4c4x" event={"ID":"e10e6f00-e11c-4f35-abad-7127db0c7cf3","Type":"ContainerDied","Data":"e5f3de3dfb82abd886df377c901d821179552144b341c3b1fa6b7f45bc7753e6"} Mar 20 16:08:04 crc kubenswrapper[4936]: I0320 16:08:04.494459 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567048-c4c4x" Mar 20 16:08:04 crc kubenswrapper[4936]: I0320 16:08:04.552073 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5x65\" (UniqueName: \"kubernetes.io/projected/e10e6f00-e11c-4f35-abad-7127db0c7cf3-kube-api-access-w5x65\") pod \"e10e6f00-e11c-4f35-abad-7127db0c7cf3\" (UID: \"e10e6f00-e11c-4f35-abad-7127db0c7cf3\") " Mar 20 16:08:04 crc kubenswrapper[4936]: I0320 16:08:04.562282 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e10e6f00-e11c-4f35-abad-7127db0c7cf3-kube-api-access-w5x65" (OuterVolumeSpecName: "kube-api-access-w5x65") pod "e10e6f00-e11c-4f35-abad-7127db0c7cf3" (UID: "e10e6f00-e11c-4f35-abad-7127db0c7cf3"). InnerVolumeSpecName "kube-api-access-w5x65". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:08:04 crc kubenswrapper[4936]: I0320 16:08:04.654211 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5x65\" (UniqueName: \"kubernetes.io/projected/e10e6f00-e11c-4f35-abad-7127db0c7cf3-kube-api-access-w5x65\") on node \"crc\" DevicePath \"\"" Mar 20 16:08:05 crc kubenswrapper[4936]: I0320 16:08:05.231496 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567048-c4c4x" event={"ID":"e10e6f00-e11c-4f35-abad-7127db0c7cf3","Type":"ContainerDied","Data":"e42a3bd0662758e76439d2f4ea698f43efa9368fbd6551c77c7e48336e5baf97"} Mar 20 16:08:05 crc kubenswrapper[4936]: I0320 16:08:05.231537 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e42a3bd0662758e76439d2f4ea698f43efa9368fbd6551c77c7e48336e5baf97" Mar 20 16:08:05 crc kubenswrapper[4936]: I0320 16:08:05.231576 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567048-c4c4x" Mar 20 16:08:28 crc kubenswrapper[4936]: I0320 16:08:28.854818 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:08:28 crc kubenswrapper[4936]: I0320 16:08:28.855470 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:08:58 crc kubenswrapper[4936]: I0320 16:08:58.854753 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:08:58 crc kubenswrapper[4936]: I0320 16:08:58.855310 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:09:28 crc kubenswrapper[4936]: I0320 16:09:28.855255 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:09:28 crc kubenswrapper[4936]: I0320 16:09:28.855960 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:09:28 crc kubenswrapper[4936]: I0320 16:09:28.856028 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:09:28 crc kubenswrapper[4936]: I0320 16:09:28.857403 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ad7190bb15359c2cb63ff782fee447cfc84994fffc54cec84eef49413cfae9d9"} pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 16:09:28 crc kubenswrapper[4936]: I0320 16:09:28.857736 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" containerID="cri-o://ad7190bb15359c2cb63ff782fee447cfc84994fffc54cec84eef49413cfae9d9" gracePeriod=600 Mar 20 16:09:29 crc kubenswrapper[4936]: I0320 16:09:29.824920 4936 generic.go:334] "Generic (PLEG): container finished" podID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerID="ad7190bb15359c2cb63ff782fee447cfc84994fffc54cec84eef49413cfae9d9" exitCode=0 Mar 20 16:09:29 crc kubenswrapper[4936]: I0320 16:09:29.825064 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerDied","Data":"ad7190bb15359c2cb63ff782fee447cfc84994fffc54cec84eef49413cfae9d9"} Mar 20 16:09:29 crc kubenswrapper[4936]: I0320 16:09:29.825794 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"6e12ccb52ca45cab384d2f9173c7abe5ae747d569fea13ff0f7cd5e5e7c20df0"} Mar 20 16:09:29 crc kubenswrapper[4936]: I0320 16:09:29.825829 4936 scope.go:117] "RemoveContainer" containerID="548766dd512244399f51b55b943c4d13acad4bf13ce80ec19f3b99a712997270" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.154372 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567050-4qjb9"] Mar 20 16:10:00 crc kubenswrapper[4936]: E0320 16:10:00.155602 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10e6f00-e11c-4f35-abad-7127db0c7cf3" containerName="oc" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.155628 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10e6f00-e11c-4f35-abad-7127db0c7cf3" containerName="oc" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.155854 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10e6f00-e11c-4f35-abad-7127db0c7cf3" containerName="oc" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.156398 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567050-4qjb9" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.161015 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.161349 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.166305 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.178119 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567050-4qjb9"] Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.222212 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znvj4\" (UniqueName: \"kubernetes.io/projected/ec3d876a-aacb-4ba2-93cd-9844aa1f70d9-kube-api-access-znvj4\") pod \"auto-csr-approver-29567050-4qjb9\" (UID: \"ec3d876a-aacb-4ba2-93cd-9844aa1f70d9\") " pod="openshift-infra/auto-csr-approver-29567050-4qjb9" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.323493 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znvj4\" (UniqueName: \"kubernetes.io/projected/ec3d876a-aacb-4ba2-93cd-9844aa1f70d9-kube-api-access-znvj4\") pod \"auto-csr-approver-29567050-4qjb9\" (UID: \"ec3d876a-aacb-4ba2-93cd-9844aa1f70d9\") " pod="openshift-infra/auto-csr-approver-29567050-4qjb9" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.359238 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znvj4\" (UniqueName: \"kubernetes.io/projected/ec3d876a-aacb-4ba2-93cd-9844aa1f70d9-kube-api-access-znvj4\") pod \"auto-csr-approver-29567050-4qjb9\" (UID: \"ec3d876a-aacb-4ba2-93cd-9844aa1f70d9\") " pod="openshift-infra/auto-csr-approver-29567050-4qjb9" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.517259 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567050-4qjb9" Mar 20 16:10:00 crc kubenswrapper[4936]: I0320 16:10:00.774164 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567050-4qjb9"] Mar 20 16:10:01 crc kubenswrapper[4936]: I0320 16:10:01.201776 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567050-4qjb9" event={"ID":"ec3d876a-aacb-4ba2-93cd-9844aa1f70d9","Type":"ContainerStarted","Data":"7aa2878ca23be415e2830389f23daacc8bf68cf20fd75f33a2e1c3e263583949"} Mar 20 16:10:07 crc kubenswrapper[4936]: I0320 16:10:07.251200 4936 generic.go:334] "Generic (PLEG): container finished" podID="ec3d876a-aacb-4ba2-93cd-9844aa1f70d9" containerID="03941498646bcea4ff0136be0cbd242cf6896c3efaada8cca8ad86b0b7e197c6" exitCode=0 Mar 20 16:10:07 crc kubenswrapper[4936]: I0320 16:10:07.251299 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567050-4qjb9" event={"ID":"ec3d876a-aacb-4ba2-93cd-9844aa1f70d9","Type":"ContainerDied","Data":"03941498646bcea4ff0136be0cbd242cf6896c3efaada8cca8ad86b0b7e197c6"} Mar 20 16:10:08 crc kubenswrapper[4936]: I0320 16:10:08.575568 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567050-4qjb9" Mar 20 16:10:08 crc kubenswrapper[4936]: I0320 16:10:08.637715 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znvj4\" (UniqueName: \"kubernetes.io/projected/ec3d876a-aacb-4ba2-93cd-9844aa1f70d9-kube-api-access-znvj4\") pod \"ec3d876a-aacb-4ba2-93cd-9844aa1f70d9\" (UID: \"ec3d876a-aacb-4ba2-93cd-9844aa1f70d9\") " Mar 20 16:10:08 crc kubenswrapper[4936]: I0320 16:10:08.653795 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec3d876a-aacb-4ba2-93cd-9844aa1f70d9-kube-api-access-znvj4" (OuterVolumeSpecName: "kube-api-access-znvj4") pod "ec3d876a-aacb-4ba2-93cd-9844aa1f70d9" (UID: "ec3d876a-aacb-4ba2-93cd-9844aa1f70d9"). InnerVolumeSpecName "kube-api-access-znvj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:10:08 crc kubenswrapper[4936]: I0320 16:10:08.739428 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znvj4\" (UniqueName: \"kubernetes.io/projected/ec3d876a-aacb-4ba2-93cd-9844aa1f70d9-kube-api-access-znvj4\") on node \"crc\" DevicePath \"\"" Mar 20 16:10:09 crc kubenswrapper[4936]: I0320 16:10:09.268598 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567050-4qjb9" event={"ID":"ec3d876a-aacb-4ba2-93cd-9844aa1f70d9","Type":"ContainerDied","Data":"7aa2878ca23be415e2830389f23daacc8bf68cf20fd75f33a2e1c3e263583949"} Mar 20 16:10:09 crc kubenswrapper[4936]: I0320 16:10:09.268631 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567050-4qjb9" Mar 20 16:10:09 crc kubenswrapper[4936]: I0320 16:10:09.268669 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7aa2878ca23be415e2830389f23daacc8bf68cf20fd75f33a2e1c3e263583949" Mar 20 16:10:09 crc kubenswrapper[4936]: I0320 16:10:09.660615 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567044-4hcwr"] Mar 20 16:10:09 crc kubenswrapper[4936]: I0320 16:10:09.666613 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567044-4hcwr"] Mar 20 16:10:09 crc kubenswrapper[4936]: I0320 16:10:09.866338 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="531e6b04-d999-4aca-ad9d-981c1db726d9" path="/var/lib/kubelet/pods/531e6b04-d999-4aca-ad9d-981c1db726d9/volumes" Mar 20 16:10:50 crc kubenswrapper[4936]: I0320 16:10:50.240387 4936 scope.go:117] "RemoveContainer" containerID="0e948211251ab8f49c63f1a5565d2cc9c5d22b946421def1f9e82877b318eaf3" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.900253 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv"] Mar 20 16:11:45 crc kubenswrapper[4936]: E0320 16:11:45.903060 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec3d876a-aacb-4ba2-93cd-9844aa1f70d9" containerName="oc" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.903177 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec3d876a-aacb-4ba2-93cd-9844aa1f70d9" containerName="oc" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.903354 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec3d876a-aacb-4ba2-93cd-9844aa1f70d9" containerName="oc" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.903973 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.912757 4936 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-q27kh" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.913015 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.913071 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.919679 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-lm2fv"] Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.921560 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv"] Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.921768 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-lm2fv" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.924041 4936 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-v92xb" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.929727 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-z9dsw"] Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.934654 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.937301 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-z9dsw"] Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.938377 4936 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-zqkzn" Mar 20 16:11:45 crc kubenswrapper[4936]: I0320 16:11:45.940108 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-lm2fv"] Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.095429 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qc9v\" (UniqueName: \"kubernetes.io/projected/0f954785-a2e7-46c3-9958-d243fa27da2c-kube-api-access-6qc9v\") pod \"cert-manager-cainjector-cf98fcc89-kcmnv\" (UID: \"0f954785-a2e7-46c3-9958-d243fa27da2c\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.095489 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmpzr\" (UniqueName: \"kubernetes.io/projected/084aa53e-7882-49d2-afac-d70afccc98b8-kube-api-access-vmpzr\") pod \"cert-manager-858654f9db-lm2fv\" (UID: \"084aa53e-7882-49d2-afac-d70afccc98b8\") " pod="cert-manager/cert-manager-858654f9db-lm2fv" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.095530 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wtf9\" (UniqueName: \"kubernetes.io/projected/b04f08ab-f31c-4229-8158-a589251be69c-kube-api-access-9wtf9\") pod \"cert-manager-webhook-687f57d79b-z9dsw\" (UID: \"b04f08ab-f31c-4229-8158-a589251be69c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.197215 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qc9v\" (UniqueName: \"kubernetes.io/projected/0f954785-a2e7-46c3-9958-d243fa27da2c-kube-api-access-6qc9v\") pod \"cert-manager-cainjector-cf98fcc89-kcmnv\" (UID: \"0f954785-a2e7-46c3-9958-d243fa27da2c\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.197299 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmpzr\" (UniqueName: \"kubernetes.io/projected/084aa53e-7882-49d2-afac-d70afccc98b8-kube-api-access-vmpzr\") pod \"cert-manager-858654f9db-lm2fv\" (UID: \"084aa53e-7882-49d2-afac-d70afccc98b8\") " pod="cert-manager/cert-manager-858654f9db-lm2fv" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.197384 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wtf9\" (UniqueName: \"kubernetes.io/projected/b04f08ab-f31c-4229-8158-a589251be69c-kube-api-access-9wtf9\") pod \"cert-manager-webhook-687f57d79b-z9dsw\" (UID: \"b04f08ab-f31c-4229-8158-a589251be69c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.216332 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmpzr\" (UniqueName: \"kubernetes.io/projected/084aa53e-7882-49d2-afac-d70afccc98b8-kube-api-access-vmpzr\") pod \"cert-manager-858654f9db-lm2fv\" (UID: \"084aa53e-7882-49d2-afac-d70afccc98b8\") " pod="cert-manager/cert-manager-858654f9db-lm2fv" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.218641 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qc9v\" (UniqueName: \"kubernetes.io/projected/0f954785-a2e7-46c3-9958-d243fa27da2c-kube-api-access-6qc9v\") pod \"cert-manager-cainjector-cf98fcc89-kcmnv\" (UID: \"0f954785-a2e7-46c3-9958-d243fa27da2c\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.219080 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wtf9\" (UniqueName: \"kubernetes.io/projected/b04f08ab-f31c-4229-8158-a589251be69c-kube-api-access-9wtf9\") pod \"cert-manager-webhook-687f57d79b-z9dsw\" (UID: \"b04f08ab-f31c-4229-8158-a589251be69c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.234738 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.244865 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-lm2fv" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.252352 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.687730 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-lm2fv"] Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.697265 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-z9dsw"] Mar 20 16:11:46 crc kubenswrapper[4936]: I0320 16:11:46.701210 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv"] Mar 20 16:11:46 crc kubenswrapper[4936]: W0320 16:11:46.707159 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb04f08ab_f31c_4229_8158_a589251be69c.slice/crio-5faee300059af16813d084b4cb6a97298db93804b285675a0e6b2e3778594728 WatchSource:0}: Error finding container 5faee300059af16813d084b4cb6a97298db93804b285675a0e6b2e3778594728: Status 404 returned error can't find the container with id 5faee300059af16813d084b4cb6a97298db93804b285675a0e6b2e3778594728 Mar 20 16:11:46 crc kubenswrapper[4936]: W0320 16:11:46.714245 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f954785_a2e7_46c3_9958_d243fa27da2c.slice/crio-cd62254188630bfed6287e585ee3670eadd9cd9038587e328bb7a6998d5e7672 WatchSource:0}: Error finding container cd62254188630bfed6287e585ee3670eadd9cd9038587e328bb7a6998d5e7672: Status 404 returned error can't find the container with id cd62254188630bfed6287e585ee3670eadd9cd9038587e328bb7a6998d5e7672 Mar 20 16:11:47 crc kubenswrapper[4936]: I0320 16:11:47.184089 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" event={"ID":"b04f08ab-f31c-4229-8158-a589251be69c","Type":"ContainerStarted","Data":"5faee300059af16813d084b4cb6a97298db93804b285675a0e6b2e3778594728"} Mar 20 16:11:47 crc kubenswrapper[4936]: I0320 16:11:47.185770 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv" event={"ID":"0f954785-a2e7-46c3-9958-d243fa27da2c","Type":"ContainerStarted","Data":"cd62254188630bfed6287e585ee3670eadd9cd9038587e328bb7a6998d5e7672"} Mar 20 16:11:47 crc kubenswrapper[4936]: I0320 16:11:47.188437 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-lm2fv" event={"ID":"084aa53e-7882-49d2-afac-d70afccc98b8","Type":"ContainerStarted","Data":"70091c5fbc985640f90615025b86f5f6cd43fa360793e683c15ab5fcd0c49f49"} Mar 20 16:11:51 crc kubenswrapper[4936]: I0320 16:11:51.211416 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-lm2fv" event={"ID":"084aa53e-7882-49d2-afac-d70afccc98b8","Type":"ContainerStarted","Data":"08e1b8d323c89372f0b27dbe42c1abd5281d5152765ae1da11b83e1a0f0c5733"} Mar 20 16:11:51 crc kubenswrapper[4936]: I0320 16:11:51.212678 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" event={"ID":"b04f08ab-f31c-4229-8158-a589251be69c","Type":"ContainerStarted","Data":"4705a33151b46b783616a6994b2c45ca1252ad3a8fc6dde67177e72deef0302d"} Mar 20 16:11:51 crc kubenswrapper[4936]: I0320 16:11:51.212868 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" Mar 20 16:11:51 crc kubenswrapper[4936]: I0320 16:11:51.214006 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv" event={"ID":"0f954785-a2e7-46c3-9958-d243fa27da2c","Type":"ContainerStarted","Data":"dc7853212b82100f4e83607175de2bebbd09d1d9ad5baf4fb338089a20b4a511"} Mar 20 16:11:51 crc kubenswrapper[4936]: I0320 16:11:51.227623 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-lm2fv" podStartSLOduration=2.316862542 podStartE2EDuration="6.227602841s" podCreationTimestamp="2026-03-20 16:11:45 +0000 UTC" firstStartedPulling="2026-03-20 16:11:46.698180697 +0000 UTC m=+657.644548512" lastFinishedPulling="2026-03-20 16:11:50.608920986 +0000 UTC m=+661.555288811" observedRunningTime="2026-03-20 16:11:51.224752587 +0000 UTC m=+662.171120402" watchObservedRunningTime="2026-03-20 16:11:51.227602841 +0000 UTC m=+662.173970656" Mar 20 16:11:51 crc kubenswrapper[4936]: I0320 16:11:51.257145 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-kcmnv" podStartSLOduration=2.366771554 podStartE2EDuration="6.25708081s" podCreationTimestamp="2026-03-20 16:11:45 +0000 UTC" firstStartedPulling="2026-03-20 16:11:46.72643879 +0000 UTC m=+657.672806605" lastFinishedPulling="2026-03-20 16:11:50.616748036 +0000 UTC m=+661.563115861" observedRunningTime="2026-03-20 16:11:51.246475097 +0000 UTC m=+662.192842952" watchObservedRunningTime="2026-03-20 16:11:51.25708081 +0000 UTC m=+662.203448655" Mar 20 16:11:51 crc kubenswrapper[4936]: I0320 16:11:51.278322 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" podStartSLOduration=2.358068197 podStartE2EDuration="6.278299286s" podCreationTimestamp="2026-03-20 16:11:45 +0000 UTC" firstStartedPulling="2026-03-20 16:11:46.708950164 +0000 UTC m=+657.655317979" lastFinishedPulling="2026-03-20 16:11:50.629181213 +0000 UTC m=+661.575549068" observedRunningTime="2026-03-20 16:11:51.27403076 +0000 UTC m=+662.220398585" watchObservedRunningTime="2026-03-20 16:11:51.278299286 +0000 UTC m=+662.224667101" Mar 20 16:11:55 crc kubenswrapper[4936]: I0320 16:11:55.749268 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jr675"] Mar 20 16:11:55 crc kubenswrapper[4936]: I0320 16:11:55.750825 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovn-controller" containerID="cri-o://6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83" gracePeriod=30 Mar 20 16:11:55 crc kubenswrapper[4936]: I0320 16:11:55.750891 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="sbdb" containerID="cri-o://8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670" gracePeriod=30 Mar 20 16:11:55 crc kubenswrapper[4936]: I0320 16:11:55.750992 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kube-rbac-proxy-node" containerID="cri-o://0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424" gracePeriod=30 Mar 20 16:11:55 crc kubenswrapper[4936]: I0320 16:11:55.750993 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovn-acl-logging" containerID="cri-o://010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f" gracePeriod=30 Mar 20 16:11:55 crc kubenswrapper[4936]: I0320 16:11:55.750997 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="nbdb" containerID="cri-o://9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa" gracePeriod=30 Mar 20 16:11:55 crc kubenswrapper[4936]: I0320 16:11:55.750978 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="northd" containerID="cri-o://c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa" gracePeriod=30 Mar 20 16:11:55 crc kubenswrapper[4936]: I0320 16:11:55.751146 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a" gracePeriod=30 Mar 20 16:11:55 crc kubenswrapper[4936]: I0320 16:11:55.792337 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovnkube-controller" containerID="cri-o://53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50" gracePeriod=30 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.064411 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jr675_cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736/ovn-acl-logging/0.log" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.065624 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jr675_cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736/ovn-controller/0.log" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.066376 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137558 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qfl7g"] Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.137845 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="northd" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137860 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="northd" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.137871 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kubecfg-setup" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137879 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kubecfg-setup" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.137886 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137892 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.137901 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovn-controller" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137907 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovn-controller" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.137916 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="sbdb" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137921 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="sbdb" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.137931 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kube-rbac-proxy-node" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137937 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kube-rbac-proxy-node" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.137946 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="nbdb" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137951 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="nbdb" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.137959 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovnkube-controller" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137965 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovnkube-controller" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.137971 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovn-acl-logging" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.137978 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovn-acl-logging" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.138071 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="northd" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.138084 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="nbdb" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.138095 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovn-controller" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.138105 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovn-acl-logging" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.138112 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.138132 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="ovnkube-controller" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.138141 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="kube-rbac-proxy-node" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.138151 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerName="sbdb" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.143922 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235454 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-var-lib-openvswitch\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235518 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-node-log\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235586 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkc42\" (UniqueName: \"kubernetes.io/projected/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-kube-api-access-lkc42\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235629 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-script-lib\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235654 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-netns\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235702 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-systemd-units\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235749 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovn-node-metrics-cert\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235833 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-kubelet\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235833 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235852 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-node-log" (OuterVolumeSpecName: "node-log") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235918 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235909 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235946 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.235990 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-openvswitch\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236082 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236240 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-systemd\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236459 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236565 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-log-socket\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236596 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-log-socket" (OuterVolumeSpecName: "log-socket") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236625 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-slash\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236666 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-env-overrides\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236713 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-slash" (OuterVolumeSpecName: "host-slash") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236729 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-netd\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236778 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-bin\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236809 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236813 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-var-lib-cni-networks-ovn-kubernetes\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236836 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236864 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-ovn-kubernetes\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236901 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-etc-openvswitch\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236892 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236974 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-ovn\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.236999 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237027 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-config\") pod \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\" (UID: \"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736\") " Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237011 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237048 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237061 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237213 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-slash\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237284 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-node-log\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237317 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-log-socket\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237328 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237350 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-kubelet\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237383 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-run-openvswitch\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237436 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-etc-openvswitch\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237473 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1e2beb7f-48d6-4780-b65d-ed4f86505aff-env-overrides\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237517 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-cni-bin\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237608 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zkr4\" (UniqueName: \"kubernetes.io/projected/1e2beb7f-48d6-4780-b65d-ed4f86505aff-kube-api-access-4zkr4\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237662 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-var-lib-openvswitch\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237706 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-run-ovn-kubernetes\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237746 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1e2beb7f-48d6-4780-b65d-ed4f86505aff-ovnkube-script-lib\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237791 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-cni-netd\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237854 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-systemd-units\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237891 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-run-systemd\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237921 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1e2beb7f-48d6-4780-b65d-ed4f86505aff-ovn-node-metrics-cert\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.237966 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-run-ovn\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238151 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-run-netns\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238256 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1e2beb7f-48d6-4780-b65d-ed4f86505aff-ovnkube-config\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238288 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238363 4936 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238532 4936 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238575 4936 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238593 4936 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238609 4936 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238631 4936 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-log-socket\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238648 4936 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-slash\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238665 4936 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238682 4936 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238699 4936 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238719 4936 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238739 4936 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238756 4936 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238775 4936 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238792 4936 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238809 4936 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.238824 4936 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-node-log\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.241290 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.241721 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-kube-api-access-lkc42" (OuterVolumeSpecName: "kube-api-access-lkc42") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "kube-api-access-lkc42". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.248913 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" (UID: "cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.249183 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jr675_cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736/ovn-acl-logging/0.log" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.249979 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jr675_cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736/ovn-controller/0.log" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250436 4936 generic.go:334] "Generic (PLEG): container finished" podID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerID="53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50" exitCode=0 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250468 4936 generic.go:334] "Generic (PLEG): container finished" podID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerID="8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670" exitCode=0 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250500 4936 generic.go:334] "Generic (PLEG): container finished" podID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerID="9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa" exitCode=0 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250509 4936 generic.go:334] "Generic (PLEG): container finished" podID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerID="c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa" exitCode=0 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250520 4936 generic.go:334] "Generic (PLEG): container finished" podID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerID="7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a" exitCode=0 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250529 4936 generic.go:334] "Generic (PLEG): container finished" podID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerID="0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424" exitCode=0 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250508 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250591 4936 generic.go:334] "Generic (PLEG): container finished" podID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerID="010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f" exitCode=143 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250607 4936 generic.go:334] "Generic (PLEG): container finished" podID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" containerID="6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83" exitCode=143 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250607 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250596 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250609 4936 scope.go:117] "RemoveContainer" containerID="53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.250671 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251021 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251039 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251056 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251070 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251082 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251089 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251098 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251109 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251117 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251124 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251131 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251139 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251147 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251155 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251163 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251170 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251180 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251191 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251199 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251206 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251275 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251286 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251293 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251300 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251307 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251313 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251325 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jr675" event={"ID":"cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736","Type":"ContainerDied","Data":"c96029774b5855d58e60bfc1f9bcfbf99acffaec31a6a5f3bfb970ad9a0520e4"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251337 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251345 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251353 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251359 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251366 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251372 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251378 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251385 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.251391 4936 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.254496 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zjjnf_6f8c1cbb-9378-440e-87df-1bc4e70bcaeb/kube-multus/0.log" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.254561 4936 generic.go:334] "Generic (PLEG): container finished" podID="6f8c1cbb-9378-440e-87df-1bc4e70bcaeb" containerID="4f46be2ad28273415132ddc59b7d8ad159e695efb3ac4b6500af34424f6776b8" exitCode=2 Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.254576 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zjjnf" event={"ID":"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb","Type":"ContainerDied","Data":"4f46be2ad28273415132ddc59b7d8ad159e695efb3ac4b6500af34424f6776b8"} Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.255210 4936 scope.go:117] "RemoveContainer" containerID="4f46be2ad28273415132ddc59b7d8ad159e695efb3ac4b6500af34424f6776b8" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.256065 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-z9dsw" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.283798 4936 scope.go:117] "RemoveContainer" containerID="8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.306768 4936 scope.go:117] "RemoveContainer" containerID="9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.325892 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jr675"] Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.331915 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jr675"] Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340078 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-var-lib-openvswitch\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340116 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-run-ovn-kubernetes\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340145 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1e2beb7f-48d6-4780-b65d-ed4f86505aff-ovnkube-script-lib\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340162 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-var-lib-openvswitch\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340187 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-cni-netd\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340231 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-cni-netd\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340234 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-systemd-units\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340276 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-run-systemd\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340301 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1e2beb7f-48d6-4780-b65d-ed4f86505aff-ovn-node-metrics-cert\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340303 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-systemd-units\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340326 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-run-ovn\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340351 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-run-systemd\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340368 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-run-netns\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340391 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-run-ovn\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340422 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-run-netns\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340430 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1e2beb7f-48d6-4780-b65d-ed4f86505aff-ovnkube-config\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340457 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-run-ovn-kubernetes\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340460 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340487 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340581 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-slash\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340608 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-slash\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340670 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-log-socket\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340698 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-node-log\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340788 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-node-log\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340753 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-log-socket\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340791 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-kubelet\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340828 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-kubelet\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340842 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-run-openvswitch\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340873 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-run-openvswitch\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340894 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-etc-openvswitch\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340928 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1e2beb7f-48d6-4780-b65d-ed4f86505aff-env-overrides\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340979 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-cni-bin\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.340989 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-etc-openvswitch\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.341055 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zkr4\" (UniqueName: \"kubernetes.io/projected/1e2beb7f-48d6-4780-b65d-ed4f86505aff-kube-api-access-4zkr4\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.341062 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1e2beb7f-48d6-4780-b65d-ed4f86505aff-ovnkube-script-lib\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.341085 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1e2beb7f-48d6-4780-b65d-ed4f86505aff-ovnkube-config\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.341170 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e2beb7f-48d6-4780-b65d-ed4f86505aff-host-cni-bin\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.341266 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkc42\" (UniqueName: \"kubernetes.io/projected/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-kube-api-access-lkc42\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.341281 4936 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.341293 4936 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.343165 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1e2beb7f-48d6-4780-b65d-ed4f86505aff-env-overrides\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.346224 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1e2beb7f-48d6-4780-b65d-ed4f86505aff-ovn-node-metrics-cert\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.351405 4936 scope.go:117] "RemoveContainer" containerID="c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.355788 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zkr4\" (UniqueName: \"kubernetes.io/projected/1e2beb7f-48d6-4780-b65d-ed4f86505aff-kube-api-access-4zkr4\") pod \"ovnkube-node-qfl7g\" (UID: \"1e2beb7f-48d6-4780-b65d-ed4f86505aff\") " pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.370379 4936 scope.go:117] "RemoveContainer" containerID="7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.383432 4936 scope.go:117] "RemoveContainer" containerID="0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.404559 4936 scope.go:117] "RemoveContainer" containerID="010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.421531 4936 scope.go:117] "RemoveContainer" containerID="6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.436501 4936 scope.go:117] "RemoveContainer" containerID="00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.461699 4936 scope.go:117] "RemoveContainer" containerID="53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.462202 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": container with ID starting with 53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50 not found: ID does not exist" containerID="53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.462246 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} err="failed to get container status \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": rpc error: code = NotFound desc = could not find container \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": container with ID starting with 53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.462270 4936 scope.go:117] "RemoveContainer" containerID="8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.462761 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": container with ID starting with 8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670 not found: ID does not exist" containerID="8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.462790 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} err="failed to get container status \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": rpc error: code = NotFound desc = could not find container \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": container with ID starting with 8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.462831 4936 scope.go:117] "RemoveContainer" containerID="9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.463107 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": container with ID starting with 9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa not found: ID does not exist" containerID="9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.463136 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} err="failed to get container status \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": rpc error: code = NotFound desc = could not find container \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": container with ID starting with 9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.463149 4936 scope.go:117] "RemoveContainer" containerID="c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.463500 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": container with ID starting with c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa not found: ID does not exist" containerID="c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.463530 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} err="failed to get container status \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": rpc error: code = NotFound desc = could not find container \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": container with ID starting with c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.463570 4936 scope.go:117] "RemoveContainer" containerID="7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.464208 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": container with ID starting with 7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a not found: ID does not exist" containerID="7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.464262 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} err="failed to get container status \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": rpc error: code = NotFound desc = could not find container \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": container with ID starting with 7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.464295 4936 scope.go:117] "RemoveContainer" containerID="0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.464729 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": container with ID starting with 0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424 not found: ID does not exist" containerID="0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.464755 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} err="failed to get container status \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": rpc error: code = NotFound desc = could not find container \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": container with ID starting with 0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.464770 4936 scope.go:117] "RemoveContainer" containerID="010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.465143 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f\": container with ID starting with 010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f not found: ID does not exist" containerID="010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.465171 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} err="failed to get container status \"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f\": rpc error: code = NotFound desc = could not find container \"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f\": container with ID starting with 010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.465186 4936 scope.go:117] "RemoveContainer" containerID="6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.465192 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.465509 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83\": container with ID starting with 6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83 not found: ID does not exist" containerID="6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.465588 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} err="failed to get container status \"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83\": rpc error: code = NotFound desc = could not find container \"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83\": container with ID starting with 6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.465618 4936 scope.go:117] "RemoveContainer" containerID="00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3" Mar 20 16:11:56 crc kubenswrapper[4936]: E0320 16:11:56.466018 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3\": container with ID starting with 00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3 not found: ID does not exist" containerID="00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.466040 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3"} err="failed to get container status \"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3\": rpc error: code = NotFound desc = could not find container \"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3\": container with ID starting with 00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.466056 4936 scope.go:117] "RemoveContainer" containerID="53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.466759 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} err="failed to get container status \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": rpc error: code = NotFound desc = could not find container \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": container with ID starting with 53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.466780 4936 scope.go:117] "RemoveContainer" containerID="8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.467038 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} err="failed to get container status \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": rpc error: code = NotFound desc = could not find container \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": container with ID starting with 8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.467064 4936 scope.go:117] "RemoveContainer" containerID="9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.467448 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} err="failed to get container status \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": rpc error: code = NotFound desc = could not find container \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": container with ID starting with 9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.467479 4936 scope.go:117] "RemoveContainer" containerID="c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.467942 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} err="failed to get container status \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": rpc error: code = NotFound desc = could not find container \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": container with ID starting with c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.467986 4936 scope.go:117] "RemoveContainer" containerID="7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.468313 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} err="failed to get container status \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": rpc error: code = NotFound desc = could not find container \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": container with ID starting with 7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.468333 4936 scope.go:117] "RemoveContainer" containerID="0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.468645 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} err="failed to get container status \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": rpc error: code = NotFound desc = could not find container \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": container with ID starting with 0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.468671 4936 scope.go:117] "RemoveContainer" containerID="010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.469025 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} err="failed to get container status \"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f\": rpc error: code = NotFound desc = could not find container \"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f\": container with ID starting with 010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.469066 4936 scope.go:117] "RemoveContainer" containerID="6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.469581 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} err="failed to get container status \"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83\": rpc error: code = NotFound desc = could not find container \"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83\": container with ID starting with 6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.469613 4936 scope.go:117] "RemoveContainer" containerID="00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.469915 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3"} err="failed to get container status \"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3\": rpc error: code = NotFound desc = could not find container \"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3\": container with ID starting with 00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.469975 4936 scope.go:117] "RemoveContainer" containerID="53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.471107 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} err="failed to get container status \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": rpc error: code = NotFound desc = could not find container \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": container with ID starting with 53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.471146 4936 scope.go:117] "RemoveContainer" containerID="8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.471746 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} err="failed to get container status \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": rpc error: code = NotFound desc = could not find container \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": container with ID starting with 8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.471774 4936 scope.go:117] "RemoveContainer" containerID="9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.472430 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} err="failed to get container status \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": rpc error: code = NotFound desc = could not find container \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": container with ID starting with 9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.472459 4936 scope.go:117] "RemoveContainer" containerID="c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.473284 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} err="failed to get container status \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": rpc error: code = NotFound desc = could not find container \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": container with ID starting with c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.473306 4936 scope.go:117] "RemoveContainer" containerID="7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.473536 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} err="failed to get container status \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": rpc error: code = NotFound desc = could not find container \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": container with ID starting with 7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.473579 4936 scope.go:117] "RemoveContainer" containerID="0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.473809 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} err="failed to get container status \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": rpc error: code = NotFound desc = could not find container \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": container with ID starting with 0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.473832 4936 scope.go:117] "RemoveContainer" containerID="010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.474075 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} err="failed to get container status \"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f\": rpc error: code = NotFound desc = could not find container \"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f\": container with ID starting with 010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.474104 4936 scope.go:117] "RemoveContainer" containerID="6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.474562 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} err="failed to get container status \"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83\": rpc error: code = NotFound desc = could not find container \"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83\": container with ID starting with 6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.474588 4936 scope.go:117] "RemoveContainer" containerID="00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.474953 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3"} err="failed to get container status \"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3\": rpc error: code = NotFound desc = could not find container \"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3\": container with ID starting with 00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.474983 4936 scope.go:117] "RemoveContainer" containerID="53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.475322 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} err="failed to get container status \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": rpc error: code = NotFound desc = could not find container \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": container with ID starting with 53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.475345 4936 scope.go:117] "RemoveContainer" containerID="8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.475614 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} err="failed to get container status \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": rpc error: code = NotFound desc = could not find container \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": container with ID starting with 8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.475647 4936 scope.go:117] "RemoveContainer" containerID="9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.475897 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} err="failed to get container status \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": rpc error: code = NotFound desc = could not find container \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": container with ID starting with 9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.475922 4936 scope.go:117] "RemoveContainer" containerID="c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.476209 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} err="failed to get container status \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": rpc error: code = NotFound desc = could not find container \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": container with ID starting with c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.476253 4936 scope.go:117] "RemoveContainer" containerID="7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.476495 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} err="failed to get container status \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": rpc error: code = NotFound desc = could not find container \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": container with ID starting with 7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.476520 4936 scope.go:117] "RemoveContainer" containerID="0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.476790 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} err="failed to get container status \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": rpc error: code = NotFound desc = could not find container \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": container with ID starting with 0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.476818 4936 scope.go:117] "RemoveContainer" containerID="010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.477012 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f"} err="failed to get container status \"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f\": rpc error: code = NotFound desc = could not find container \"010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f\": container with ID starting with 010aeca3c0262cc3c55f1790d42dc75c8f94462f5fe688a4e04eb067bbad318f not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.477031 4936 scope.go:117] "RemoveContainer" containerID="6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.477426 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83"} err="failed to get container status \"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83\": rpc error: code = NotFound desc = could not find container \"6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83\": container with ID starting with 6f81ec0980dbd16491b8fca74a1593f4a0ebb55fcb6478428cbae718fc318d83 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.477461 4936 scope.go:117] "RemoveContainer" containerID="00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.477735 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3"} err="failed to get container status \"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3\": rpc error: code = NotFound desc = could not find container \"00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3\": container with ID starting with 00c4265d3aea3d10fd0930d618cb391297ef582256c2b171275b973de0309ed3 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.477756 4936 scope.go:117] "RemoveContainer" containerID="53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.477991 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50"} err="failed to get container status \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": rpc error: code = NotFound desc = could not find container \"53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50\": container with ID starting with 53b29e7c7d56c855ad1de4a87de7265bc138dbcdd59443a3d5340f15e3adfb50 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.478015 4936 scope.go:117] "RemoveContainer" containerID="8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.478231 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670"} err="failed to get container status \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": rpc error: code = NotFound desc = could not find container \"8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670\": container with ID starting with 8d2ebfd409f3e3ccce9a0ad8ba9e6f24f70be593ca6ee85c9a25d0d49f5a5670 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.478254 4936 scope.go:117] "RemoveContainer" containerID="9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.478591 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa"} err="failed to get container status \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": rpc error: code = NotFound desc = could not find container \"9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa\": container with ID starting with 9c821c76adb3ba79ac009210dcc0231092e849da13d3cdb29d56db258a9c6bfa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.478609 4936 scope.go:117] "RemoveContainer" containerID="c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.478858 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa"} err="failed to get container status \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": rpc error: code = NotFound desc = could not find container \"c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa\": container with ID starting with c80a85f336f6db7b18bc0128abeeac4ce69736afdb36b2c365376168e0baf8fa not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.478883 4936 scope.go:117] "RemoveContainer" containerID="7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.479124 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a"} err="failed to get container status \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": rpc error: code = NotFound desc = could not find container \"7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a\": container with ID starting with 7f815faa0a626501e38f957deda334e614a74d0c7fcb1689e725d9af29f2e78a not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.479144 4936 scope.go:117] "RemoveContainer" containerID="0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424" Mar 20 16:11:56 crc kubenswrapper[4936]: I0320 16:11:56.479388 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424"} err="failed to get container status \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": rpc error: code = NotFound desc = could not find container \"0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424\": container with ID starting with 0ac0253e572b5a195243b75e4ad09c6d7796796e7c62b9f42a6e7bedce9b6424 not found: ID does not exist" Mar 20 16:11:56 crc kubenswrapper[4936]: W0320 16:11:56.488723 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e2beb7f_48d6_4780_b65d_ed4f86505aff.slice/crio-22f34bafd32ae0f8de3a96f2f738ef13f29185aa1b4afb75af311ff316129e10 WatchSource:0}: Error finding container 22f34bafd32ae0f8de3a96f2f738ef13f29185aa1b4afb75af311ff316129e10: Status 404 returned error can't find the container with id 22f34bafd32ae0f8de3a96f2f738ef13f29185aa1b4afb75af311ff316129e10 Mar 20 16:11:57 crc kubenswrapper[4936]: I0320 16:11:57.263279 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zjjnf_6f8c1cbb-9378-440e-87df-1bc4e70bcaeb/kube-multus/0.log" Mar 20 16:11:57 crc kubenswrapper[4936]: I0320 16:11:57.263649 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zjjnf" event={"ID":"6f8c1cbb-9378-440e-87df-1bc4e70bcaeb","Type":"ContainerStarted","Data":"7c073c729cc2f65e17ef94de6c20ec3ea6d3b71727f3a0aebe6f57d16a767456"} Mar 20 16:11:57 crc kubenswrapper[4936]: I0320 16:11:57.265915 4936 generic.go:334] "Generic (PLEG): container finished" podID="1e2beb7f-48d6-4780-b65d-ed4f86505aff" containerID="d2bf7ccdfb58281a0254fc39edc12f3f8fac0570c1546764e407e10c1fdc32a0" exitCode=0 Mar 20 16:11:57 crc kubenswrapper[4936]: I0320 16:11:57.266054 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerDied","Data":"d2bf7ccdfb58281a0254fc39edc12f3f8fac0570c1546764e407e10c1fdc32a0"} Mar 20 16:11:57 crc kubenswrapper[4936]: I0320 16:11:57.266102 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerStarted","Data":"22f34bafd32ae0f8de3a96f2f738ef13f29185aa1b4afb75af311ff316129e10"} Mar 20 16:11:57 crc kubenswrapper[4936]: I0320 16:11:57.859132 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736" path="/var/lib/kubelet/pods/cc7d9e09-9223-4f3a-8a8a-1ba83aa0e736/volumes" Mar 20 16:11:58 crc kubenswrapper[4936]: I0320 16:11:58.277912 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerStarted","Data":"400e574f407e228c29dabb04598fb79a29f66f4e1b09ffc47ccb916b3cba636c"} Mar 20 16:11:58 crc kubenswrapper[4936]: I0320 16:11:58.277955 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerStarted","Data":"486082390d32a1c4aa54d54de31156770be420d9632805d094a941bfb73b9897"} Mar 20 16:11:58 crc kubenswrapper[4936]: I0320 16:11:58.277969 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerStarted","Data":"5d09b06e2a81ca6534759f14b8cfa7196fda2f29b62a586708b0d11ffbe3261e"} Mar 20 16:11:58 crc kubenswrapper[4936]: I0320 16:11:58.277981 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerStarted","Data":"e0ee905e8255125ee943989732c2e3b72c073b28f7605959eb8e8971b6a00270"} Mar 20 16:11:58 crc kubenswrapper[4936]: I0320 16:11:58.277994 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerStarted","Data":"28dc39d7e7ff40847f3e5b5f7cec16403280c6197e14a2877375a564c6988ff8"} Mar 20 16:11:58 crc kubenswrapper[4936]: I0320 16:11:58.278004 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerStarted","Data":"b7d3b5ec95e39def55ed5af615468a1a9fc40357136ede782504c0db5803994e"} Mar 20 16:11:58 crc kubenswrapper[4936]: I0320 16:11:58.856349 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:11:58 crc kubenswrapper[4936]: I0320 16:11:58.856931 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.140960 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567052-lctck"] Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.142098 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.144785 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.144926 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.145016 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.291723 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp4js\" (UniqueName: \"kubernetes.io/projected/b030e112-1fd3-4be5-a0a7-68ab85e05f93-kube-api-access-pp4js\") pod \"auto-csr-approver-29567052-lctck\" (UID: \"b030e112-1fd3-4be5-a0a7-68ab85e05f93\") " pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.296150 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerStarted","Data":"ad4038197ca1dad2a54405d8d76d753ed27040084743472a690f1f4c4d475be5"} Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.392430 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp4js\" (UniqueName: \"kubernetes.io/projected/b030e112-1fd3-4be5-a0a7-68ab85e05f93-kube-api-access-pp4js\") pod \"auto-csr-approver-29567052-lctck\" (UID: \"b030e112-1fd3-4be5-a0a7-68ab85e05f93\") " pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.424464 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp4js\" (UniqueName: \"kubernetes.io/projected/b030e112-1fd3-4be5-a0a7-68ab85e05f93-kube-api-access-pp4js\") pod \"auto-csr-approver-29567052-lctck\" (UID: \"b030e112-1fd3-4be5-a0a7-68ab85e05f93\") " pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:00 crc kubenswrapper[4936]: I0320 16:12:00.507868 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:00 crc kubenswrapper[4936]: E0320 16:12:00.544485 4936 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29567052-lctck_openshift-infra_b030e112-1fd3-4be5-a0a7-68ab85e05f93_0(bcd287f9d000dfdc01f2d0e2e8b8f4b30693ca0bea220e7048ce8a311ed962ca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 16:12:00 crc kubenswrapper[4936]: E0320 16:12:00.544627 4936 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29567052-lctck_openshift-infra_b030e112-1fd3-4be5-a0a7-68ab85e05f93_0(bcd287f9d000dfdc01f2d0e2e8b8f4b30693ca0bea220e7048ce8a311ed962ca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:00 crc kubenswrapper[4936]: E0320 16:12:00.544669 4936 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29567052-lctck_openshift-infra_b030e112-1fd3-4be5-a0a7-68ab85e05f93_0(bcd287f9d000dfdc01f2d0e2e8b8f4b30693ca0bea220e7048ce8a311ed962ca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:00 crc kubenswrapper[4936]: E0320 16:12:00.544746 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29567052-lctck_openshift-infra(b030e112-1fd3-4be5-a0a7-68ab85e05f93)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29567052-lctck_openshift-infra(b030e112-1fd3-4be5-a0a7-68ab85e05f93)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29567052-lctck_openshift-infra_b030e112-1fd3-4be5-a0a7-68ab85e05f93_0(bcd287f9d000dfdc01f2d0e2e8b8f4b30693ca0bea220e7048ce8a311ed962ca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29567052-lctck" podUID="b030e112-1fd3-4be5-a0a7-68ab85e05f93" Mar 20 16:12:03 crc kubenswrapper[4936]: I0320 16:12:03.345458 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" event={"ID":"1e2beb7f-48d6-4780-b65d-ed4f86505aff","Type":"ContainerStarted","Data":"93cd113ac9b95f41ccfe621d428695c65f529a4f42e5755b7643662afa013f65"} Mar 20 16:12:03 crc kubenswrapper[4936]: I0320 16:12:03.352438 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:12:03 crc kubenswrapper[4936]: I0320 16:12:03.352485 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:12:03 crc kubenswrapper[4936]: I0320 16:12:03.383120 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" podStartSLOduration=7.383090741 podStartE2EDuration="7.383090741s" podCreationTimestamp="2026-03-20 16:11:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:12:03.3800294 +0000 UTC m=+674.326397245" watchObservedRunningTime="2026-03-20 16:12:03.383090741 +0000 UTC m=+674.329458586" Mar 20 16:12:03 crc kubenswrapper[4936]: I0320 16:12:03.406043 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:12:04 crc kubenswrapper[4936]: I0320 16:12:04.341035 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567052-lctck"] Mar 20 16:12:04 crc kubenswrapper[4936]: I0320 16:12:04.341263 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:04 crc kubenswrapper[4936]: I0320 16:12:04.341988 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:04 crc kubenswrapper[4936]: I0320 16:12:04.352815 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:12:04 crc kubenswrapper[4936]: E0320 16:12:04.381984 4936 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29567052-lctck_openshift-infra_b030e112-1fd3-4be5-a0a7-68ab85e05f93_0(4cb7761d31d53b32f52cbe861ce6674268b610fc85e8f6e51693739530d5001b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 16:12:04 crc kubenswrapper[4936]: E0320 16:12:04.382079 4936 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29567052-lctck_openshift-infra_b030e112-1fd3-4be5-a0a7-68ab85e05f93_0(4cb7761d31d53b32f52cbe861ce6674268b610fc85e8f6e51693739530d5001b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:04 crc kubenswrapper[4936]: E0320 16:12:04.382117 4936 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29567052-lctck_openshift-infra_b030e112-1fd3-4be5-a0a7-68ab85e05f93_0(4cb7761d31d53b32f52cbe861ce6674268b610fc85e8f6e51693739530d5001b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:04 crc kubenswrapper[4936]: E0320 16:12:04.382185 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29567052-lctck_openshift-infra(b030e112-1fd3-4be5-a0a7-68ab85e05f93)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29567052-lctck_openshift-infra(b030e112-1fd3-4be5-a0a7-68ab85e05f93)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29567052-lctck_openshift-infra_b030e112-1fd3-4be5-a0a7-68ab85e05f93_0(4cb7761d31d53b32f52cbe861ce6674268b610fc85e8f6e51693739530d5001b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29567052-lctck" podUID="b030e112-1fd3-4be5-a0a7-68ab85e05f93" Mar 20 16:12:04 crc kubenswrapper[4936]: I0320 16:12:04.402323 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:12:16 crc kubenswrapper[4936]: I0320 16:12:16.854812 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:16 crc kubenswrapper[4936]: I0320 16:12:16.856003 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:17 crc kubenswrapper[4936]: I0320 16:12:17.341924 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567052-lctck"] Mar 20 16:12:17 crc kubenswrapper[4936]: W0320 16:12:17.353474 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb030e112_1fd3_4be5_a0a7_68ab85e05f93.slice/crio-183d1358b3371c2d67ea9f87424f5521eb8feaad1319da45d120ee902b1acfb3 WatchSource:0}: Error finding container 183d1358b3371c2d67ea9f87424f5521eb8feaad1319da45d120ee902b1acfb3: Status 404 returned error can't find the container with id 183d1358b3371c2d67ea9f87424f5521eb8feaad1319da45d120ee902b1acfb3 Mar 20 16:12:17 crc kubenswrapper[4936]: I0320 16:12:17.445651 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567052-lctck" event={"ID":"b030e112-1fd3-4be5-a0a7-68ab85e05f93","Type":"ContainerStarted","Data":"183d1358b3371c2d67ea9f87424f5521eb8feaad1319da45d120ee902b1acfb3"} Mar 20 16:12:20 crc kubenswrapper[4936]: I0320 16:12:20.474649 4936 generic.go:334] "Generic (PLEG): container finished" podID="b030e112-1fd3-4be5-a0a7-68ab85e05f93" containerID="836dcf855452eef8c5987de7d0547f19402e63dbed8f386726c2544337c003b7" exitCode=0 Mar 20 16:12:20 crc kubenswrapper[4936]: I0320 16:12:20.474719 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567052-lctck" event={"ID":"b030e112-1fd3-4be5-a0a7-68ab85e05f93","Type":"ContainerDied","Data":"836dcf855452eef8c5987de7d0547f19402e63dbed8f386726c2544337c003b7"} Mar 20 16:12:21 crc kubenswrapper[4936]: I0320 16:12:21.763148 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:21 crc kubenswrapper[4936]: I0320 16:12:21.807937 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp4js\" (UniqueName: \"kubernetes.io/projected/b030e112-1fd3-4be5-a0a7-68ab85e05f93-kube-api-access-pp4js\") pod \"b030e112-1fd3-4be5-a0a7-68ab85e05f93\" (UID: \"b030e112-1fd3-4be5-a0a7-68ab85e05f93\") " Mar 20 16:12:21 crc kubenswrapper[4936]: I0320 16:12:21.816535 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b030e112-1fd3-4be5-a0a7-68ab85e05f93-kube-api-access-pp4js" (OuterVolumeSpecName: "kube-api-access-pp4js") pod "b030e112-1fd3-4be5-a0a7-68ab85e05f93" (UID: "b030e112-1fd3-4be5-a0a7-68ab85e05f93"). InnerVolumeSpecName "kube-api-access-pp4js". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:12:21 crc kubenswrapper[4936]: I0320 16:12:21.910771 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp4js\" (UniqueName: \"kubernetes.io/projected/b030e112-1fd3-4be5-a0a7-68ab85e05f93-kube-api-access-pp4js\") on node \"crc\" DevicePath \"\"" Mar 20 16:12:22 crc kubenswrapper[4936]: I0320 16:12:22.493411 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567052-lctck" event={"ID":"b030e112-1fd3-4be5-a0a7-68ab85e05f93","Type":"ContainerDied","Data":"183d1358b3371c2d67ea9f87424f5521eb8feaad1319da45d120ee902b1acfb3"} Mar 20 16:12:22 crc kubenswrapper[4936]: I0320 16:12:22.493633 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="183d1358b3371c2d67ea9f87424f5521eb8feaad1319da45d120ee902b1acfb3" Mar 20 16:12:22 crc kubenswrapper[4936]: I0320 16:12:22.493574 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567052-lctck" Mar 20 16:12:22 crc kubenswrapper[4936]: I0320 16:12:22.841432 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567046-nr4fn"] Mar 20 16:12:22 crc kubenswrapper[4936]: I0320 16:12:22.847756 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567046-nr4fn"] Mar 20 16:12:23 crc kubenswrapper[4936]: I0320 16:12:23.866942 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1ffe995-5ea7-46df-9389-1ac1e6bd729b" path="/var/lib/kubelet/pods/a1ffe995-5ea7-46df-9389-1ac1e6bd729b/volumes" Mar 20 16:12:26 crc kubenswrapper[4936]: I0320 16:12:26.486035 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qfl7g" Mar 20 16:12:28 crc kubenswrapper[4936]: I0320 16:12:28.854924 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:12:28 crc kubenswrapper[4936]: I0320 16:12:28.855250 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.657024 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm"] Mar 20 16:12:34 crc kubenswrapper[4936]: E0320 16:12:34.657718 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b030e112-1fd3-4be5-a0a7-68ab85e05f93" containerName="oc" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.657731 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b030e112-1fd3-4be5-a0a7-68ab85e05f93" containerName="oc" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.657824 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="b030e112-1fd3-4be5-a0a7-68ab85e05f93" containerName="oc" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.658507 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.659927 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.665361 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm"] Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.771981 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.772248 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2t7t\" (UniqueName: \"kubernetes.io/projected/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-kube-api-access-f2t7t\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.772350 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.874229 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2t7t\" (UniqueName: \"kubernetes.io/projected/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-kube-api-access-f2t7t\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.874299 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.874334 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.874884 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.875215 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:34 crc kubenswrapper[4936]: I0320 16:12:34.901019 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2t7t\" (UniqueName: \"kubernetes.io/projected/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-kube-api-access-f2t7t\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:35 crc kubenswrapper[4936]: I0320 16:12:35.006299 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:35 crc kubenswrapper[4936]: I0320 16:12:35.413492 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm"] Mar 20 16:12:35 crc kubenswrapper[4936]: I0320 16:12:35.567804 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" event={"ID":"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a","Type":"ContainerStarted","Data":"a466826ababcb98f5084085e136864bec049bcff0104e667e06d1b91abb9c3b6"} Mar 20 16:12:35 crc kubenswrapper[4936]: I0320 16:12:35.568162 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" event={"ID":"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a","Type":"ContainerStarted","Data":"52934441b4d4c352fe8ef9ed1a05b20f8da5234ea8f062f35d97e520133740e9"} Mar 20 16:12:36 crc kubenswrapper[4936]: I0320 16:12:36.574166 4936 generic.go:334] "Generic (PLEG): container finished" podID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerID="a466826ababcb98f5084085e136864bec049bcff0104e667e06d1b91abb9c3b6" exitCode=0 Mar 20 16:12:36 crc kubenswrapper[4936]: I0320 16:12:36.574215 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" event={"ID":"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a","Type":"ContainerDied","Data":"a466826ababcb98f5084085e136864bec049bcff0104e667e06d1b91abb9c3b6"} Mar 20 16:12:38 crc kubenswrapper[4936]: I0320 16:12:38.588004 4936 generic.go:334] "Generic (PLEG): container finished" podID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerID="e8103541d0fd56225d44c251f91180318585aff2631ac85785179188e78f526a" exitCode=0 Mar 20 16:12:38 crc kubenswrapper[4936]: I0320 16:12:38.588102 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" event={"ID":"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a","Type":"ContainerDied","Data":"e8103541d0fd56225d44c251f91180318585aff2631ac85785179188e78f526a"} Mar 20 16:12:39 crc kubenswrapper[4936]: I0320 16:12:39.598688 4936 generic.go:334] "Generic (PLEG): container finished" podID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerID="fb03a7f9c1b119648678350e24bc859f1402ad3660305510cc983e886e8151aa" exitCode=0 Mar 20 16:12:39 crc kubenswrapper[4936]: I0320 16:12:39.598831 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" event={"ID":"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a","Type":"ContainerDied","Data":"fb03a7f9c1b119648678350e24bc859f1402ad3660305510cc983e886e8151aa"} Mar 20 16:12:40 crc kubenswrapper[4936]: I0320 16:12:40.885359 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.068647 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2t7t\" (UniqueName: \"kubernetes.io/projected/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-kube-api-access-f2t7t\") pod \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.068818 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-util\") pod \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.068858 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-bundle\") pod \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\" (UID: \"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a\") " Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.069914 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-bundle" (OuterVolumeSpecName: "bundle") pod "88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" (UID: "88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.078777 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-util" (OuterVolumeSpecName: "util") pod "88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" (UID: "88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.085944 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-kube-api-access-f2t7t" (OuterVolumeSpecName: "kube-api-access-f2t7t") pod "88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" (UID: "88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a"). InnerVolumeSpecName "kube-api-access-f2t7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.170205 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2t7t\" (UniqueName: \"kubernetes.io/projected/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-kube-api-access-f2t7t\") on node \"crc\" DevicePath \"\"" Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.170265 4936 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-util\") on node \"crc\" DevicePath \"\"" Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.170290 4936 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.622307 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" event={"ID":"88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a","Type":"ContainerDied","Data":"52934441b4d4c352fe8ef9ed1a05b20f8da5234ea8f062f35d97e520133740e9"} Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.622384 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52934441b4d4c352fe8ef9ed1a05b20f8da5234ea8f062f35d97e520133740e9" Mar 20 16:12:41 crc kubenswrapper[4936]: I0320 16:12:41.622407 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.255252 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-b94r2"] Mar 20 16:12:46 crc kubenswrapper[4936]: E0320 16:12:46.255463 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerName="extract" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.255475 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerName="extract" Mar 20 16:12:46 crc kubenswrapper[4936]: E0320 16:12:46.255491 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerName="pull" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.255496 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerName="pull" Mar 20 16:12:46 crc kubenswrapper[4936]: E0320 16:12:46.255506 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerName="util" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.255511 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerName="util" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.255630 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a" containerName="extract" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.256020 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-b94r2" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.257920 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.258198 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-xs7kf" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.258466 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.275049 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-b94r2"] Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.352938 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dndbh\" (UniqueName: \"kubernetes.io/projected/e22309a2-eb8b-4a9e-8f56-43c169803a33-kube-api-access-dndbh\") pod \"nmstate-operator-796d4cfff4-b94r2\" (UID: \"e22309a2-eb8b-4a9e-8f56-43c169803a33\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-b94r2" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.453831 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dndbh\" (UniqueName: \"kubernetes.io/projected/e22309a2-eb8b-4a9e-8f56-43c169803a33-kube-api-access-dndbh\") pod \"nmstate-operator-796d4cfff4-b94r2\" (UID: \"e22309a2-eb8b-4a9e-8f56-43c169803a33\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-b94r2" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.479853 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dndbh\" (UniqueName: \"kubernetes.io/projected/e22309a2-eb8b-4a9e-8f56-43c169803a33-kube-api-access-dndbh\") pod \"nmstate-operator-796d4cfff4-b94r2\" (UID: \"e22309a2-eb8b-4a9e-8f56-43c169803a33\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-b94r2" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.582072 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-b94r2" Mar 20 16:12:46 crc kubenswrapper[4936]: I0320 16:12:46.847864 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-b94r2"] Mar 20 16:12:47 crc kubenswrapper[4936]: I0320 16:12:47.664006 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-b94r2" event={"ID":"e22309a2-eb8b-4a9e-8f56-43c169803a33","Type":"ContainerStarted","Data":"fa3090c3858bbf2597c42a1fc13733bf8d77f854c0dde2db56e81220542aa086"} Mar 20 16:12:50 crc kubenswrapper[4936]: I0320 16:12:50.321140 4936 scope.go:117] "RemoveContainer" containerID="e7773d332dfd7e5acd15f8dc7a3466a64370e0e72622fea11316295c4202d5d2" Mar 20 16:12:50 crc kubenswrapper[4936]: I0320 16:12:50.689264 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-b94r2" event={"ID":"e22309a2-eb8b-4a9e-8f56-43c169803a33","Type":"ContainerStarted","Data":"30ee56da6f7bd3b3a9c43633196383fc5efa982e4517661ceb5688238e67081e"} Mar 20 16:12:50 crc kubenswrapper[4936]: I0320 16:12:50.718927 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-b94r2" podStartSLOduration=1.5985606780000001 podStartE2EDuration="4.71889571s" podCreationTimestamp="2026-03-20 16:12:46 +0000 UTC" firstStartedPulling="2026-03-20 16:12:46.860759341 +0000 UTC m=+717.807127156" lastFinishedPulling="2026-03-20 16:12:49.981094373 +0000 UTC m=+720.927462188" observedRunningTime="2026-03-20 16:12:50.71583741 +0000 UTC m=+721.662205295" watchObservedRunningTime="2026-03-20 16:12:50.71889571 +0000 UTC m=+721.665263565" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.807746 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr"] Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.808599 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.810750 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.810811 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-bnlgf" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.813872 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4"] Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.815304 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.828204 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-h59z7"] Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.829032 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.836620 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4"] Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.861718 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/36c16ea1-7b8b-4175-8647-92fc61ee66d5-dbus-socket\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.861926 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8b4af217-1e38-47e1-b306-a9d4d7ed90f8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-dqbtr\" (UID: \"8b4af217-1e38-47e1-b306-a9d4d7ed90f8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.861965 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/36c16ea1-7b8b-4175-8647-92fc61ee66d5-ovs-socket\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.861992 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2gkm\" (UniqueName: \"kubernetes.io/projected/36c16ea1-7b8b-4175-8647-92fc61ee66d5-kube-api-access-f2gkm\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.862042 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27lkj\" (UniqueName: \"kubernetes.io/projected/f0f1b197-b717-44db-aed0-98ae53aba831-kube-api-access-27lkj\") pod \"nmstate-metrics-9b8c8685d-mmtw4\" (UID: \"f0f1b197-b717-44db-aed0-98ae53aba831\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.862080 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/36c16ea1-7b8b-4175-8647-92fc61ee66d5-nmstate-lock\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.862116 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbfnv\" (UniqueName: \"kubernetes.io/projected/8b4af217-1e38-47e1-b306-a9d4d7ed90f8-kube-api-access-bbfnv\") pod \"nmstate-webhook-5f558f5558-dqbtr\" (UID: \"8b4af217-1e38-47e1-b306-a9d4d7ed90f8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.892179 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr"] Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.947994 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh"] Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.948820 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:54 crc kubenswrapper[4936]: W0320 16:12:54.950822 4936 reflector.go:561] object-"openshift-nmstate"/"plugin-serving-cert": failed to list *v1.Secret: secrets "plugin-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Mar 20 16:12:54 crc kubenswrapper[4936]: E0320 16:12:54.950867 4936 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"plugin-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"plugin-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 20 16:12:54 crc kubenswrapper[4936]: W0320 16:12:54.950935 4936 reflector.go:561] object-"openshift-nmstate"/"default-dockercfg-wrdsh": failed to list *v1.Secret: secrets "default-dockercfg-wrdsh" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Mar 20 16:12:54 crc kubenswrapper[4936]: E0320 16:12:54.950955 4936 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"default-dockercfg-wrdsh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-wrdsh\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 20 16:12:54 crc kubenswrapper[4936]: W0320 16:12:54.951105 4936 reflector.go:561] object-"openshift-nmstate"/"nginx-conf": failed to list *v1.ConfigMap: configmaps "nginx-conf" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Mar 20 16:12:54 crc kubenswrapper[4936]: E0320 16:12:54.951132 4936 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"nginx-conf\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"nginx-conf\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.962663 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/36c16ea1-7b8b-4175-8647-92fc61ee66d5-dbus-socket\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.962745 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dvwg\" (UniqueName: \"kubernetes.io/projected/00920150-af2f-46d4-a3cf-bdfea2640c89-kube-api-access-5dvwg\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.962777 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8b4af217-1e38-47e1-b306-a9d4d7ed90f8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-dqbtr\" (UID: \"8b4af217-1e38-47e1-b306-a9d4d7ed90f8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.962810 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/36c16ea1-7b8b-4175-8647-92fc61ee66d5-ovs-socket\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.962836 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/00920150-af2f-46d4-a3cf-bdfea2640c89-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.962861 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2gkm\" (UniqueName: \"kubernetes.io/projected/36c16ea1-7b8b-4175-8647-92fc61ee66d5-kube-api-access-f2gkm\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.962903 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27lkj\" (UniqueName: \"kubernetes.io/projected/f0f1b197-b717-44db-aed0-98ae53aba831-kube-api-access-27lkj\") pod \"nmstate-metrics-9b8c8685d-mmtw4\" (UID: \"f0f1b197-b717-44db-aed0-98ae53aba831\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.962933 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/00920150-af2f-46d4-a3cf-bdfea2640c89-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.962971 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/36c16ea1-7b8b-4175-8647-92fc61ee66d5-nmstate-lock\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.963000 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbfnv\" (UniqueName: \"kubernetes.io/projected/8b4af217-1e38-47e1-b306-a9d4d7ed90f8-kube-api-access-bbfnv\") pod \"nmstate-webhook-5f558f5558-dqbtr\" (UID: \"8b4af217-1e38-47e1-b306-a9d4d7ed90f8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.963609 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/36c16ea1-7b8b-4175-8647-92fc61ee66d5-dbus-socket\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.964671 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/36c16ea1-7b8b-4175-8647-92fc61ee66d5-nmstate-lock\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.964720 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/36c16ea1-7b8b-4175-8647-92fc61ee66d5-ovs-socket\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.965266 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh"] Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.973401 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8b4af217-1e38-47e1-b306-a9d4d7ed90f8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-dqbtr\" (UID: \"8b4af217-1e38-47e1-b306-a9d4d7ed90f8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.980165 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27lkj\" (UniqueName: \"kubernetes.io/projected/f0f1b197-b717-44db-aed0-98ae53aba831-kube-api-access-27lkj\") pod \"nmstate-metrics-9b8c8685d-mmtw4\" (UID: \"f0f1b197-b717-44db-aed0-98ae53aba831\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.988033 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2gkm\" (UniqueName: \"kubernetes.io/projected/36c16ea1-7b8b-4175-8647-92fc61ee66d5-kube-api-access-f2gkm\") pod \"nmstate-handler-h59z7\" (UID: \"36c16ea1-7b8b-4175-8647-92fc61ee66d5\") " pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:54 crc kubenswrapper[4936]: I0320 16:12:54.995486 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbfnv\" (UniqueName: \"kubernetes.io/projected/8b4af217-1e38-47e1-b306-a9d4d7ed90f8-kube-api-access-bbfnv\") pod \"nmstate-webhook-5f558f5558-dqbtr\" (UID: \"8b4af217-1e38-47e1-b306-a9d4d7ed90f8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.067077 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/00920150-af2f-46d4-a3cf-bdfea2640c89-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.067151 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/00920150-af2f-46d4-a3cf-bdfea2640c89-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.067224 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dvwg\" (UniqueName: \"kubernetes.io/projected/00920150-af2f-46d4-a3cf-bdfea2640c89-kube-api-access-5dvwg\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.087422 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dvwg\" (UniqueName: \"kubernetes.io/projected/00920150-af2f-46d4-a3cf-bdfea2640c89-kube-api-access-5dvwg\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.127094 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.135808 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.141913 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-849fb747cf-nmlzn"] Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.142409 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.142654 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.155468 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-849fb747cf-nmlzn"] Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.168315 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-trusted-ca-bundle\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.168359 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnn2t\" (UniqueName: \"kubernetes.io/projected/16f319af-481b-4398-84e2-a6ced6cbbd0c-kube-api-access-lnn2t\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.168380 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-service-ca\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.168401 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-oauth-serving-cert\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.168525 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/16f319af-481b-4398-84e2-a6ced6cbbd0c-console-serving-cert\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.168621 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/16f319af-481b-4398-84e2-a6ced6cbbd0c-console-oauth-config\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.168671 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-console-config\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.269517 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-oauth-serving-cert\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.269865 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/16f319af-481b-4398-84e2-a6ced6cbbd0c-console-serving-cert\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.269902 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/16f319af-481b-4398-84e2-a6ced6cbbd0c-console-oauth-config\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.269938 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-console-config\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.270006 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-trusted-ca-bundle\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.270037 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnn2t\" (UniqueName: \"kubernetes.io/projected/16f319af-481b-4398-84e2-a6ced6cbbd0c-kube-api-access-lnn2t\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.270061 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-service-ca\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.271188 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-service-ca\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.271710 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-oauth-serving-cert\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.272664 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-console-config\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.276612 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16f319af-481b-4398-84e2-a6ced6cbbd0c-trusted-ca-bundle\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.278089 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/16f319af-481b-4398-84e2-a6ced6cbbd0c-console-oauth-config\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.287432 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/16f319af-481b-4398-84e2-a6ced6cbbd0c-console-serving-cert\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.295438 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnn2t\" (UniqueName: \"kubernetes.io/projected/16f319af-481b-4398-84e2-a6ced6cbbd0c-kube-api-access-lnn2t\") pod \"console-849fb747cf-nmlzn\" (UID: \"16f319af-481b-4398-84e2-a6ced6cbbd0c\") " pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.495915 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.565433 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr"] Mar 20 16:12:55 crc kubenswrapper[4936]: W0320 16:12:55.573939 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b4af217_1e38_47e1_b306_a9d4d7ed90f8.slice/crio-a652e1a3e8ca41add08f46064de8dff5242ae2f7a8c4aa264e1136eb8241c37c WatchSource:0}: Error finding container a652e1a3e8ca41add08f46064de8dff5242ae2f7a8c4aa264e1136eb8241c37c: Status 404 returned error can't find the container with id a652e1a3e8ca41add08f46064de8dff5242ae2f7a8c4aa264e1136eb8241c37c Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.635472 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4"] Mar 20 16:12:55 crc kubenswrapper[4936]: W0320 16:12:55.642507 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0f1b197_b717_44db_aed0_98ae53aba831.slice/crio-9f713bbc300ceed0b324bdc50991e520474113f0b255f95dd2f0dcf981e34a20 WatchSource:0}: Error finding container 9f713bbc300ceed0b324bdc50991e520474113f0b255f95dd2f0dcf981e34a20: Status 404 returned error can't find the container with id 9f713bbc300ceed0b324bdc50991e520474113f0b255f95dd2f0dcf981e34a20 Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.706521 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-849fb747cf-nmlzn"] Mar 20 16:12:55 crc kubenswrapper[4936]: W0320 16:12:55.708983 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16f319af_481b_4398_84e2_a6ced6cbbd0c.slice/crio-5972b8789bec01e0bddb5f1dcac4cc1322a177ce423ff7715c81619d683e35e0 WatchSource:0}: Error finding container 5972b8789bec01e0bddb5f1dcac4cc1322a177ce423ff7715c81619d683e35e0: Status 404 returned error can't find the container with id 5972b8789bec01e0bddb5f1dcac4cc1322a177ce423ff7715c81619d683e35e0 Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.716094 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4" event={"ID":"f0f1b197-b717-44db-aed0-98ae53aba831","Type":"ContainerStarted","Data":"9f713bbc300ceed0b324bdc50991e520474113f0b255f95dd2f0dcf981e34a20"} Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.716970 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-h59z7" event={"ID":"36c16ea1-7b8b-4175-8647-92fc61ee66d5","Type":"ContainerStarted","Data":"f77646d899b11af76da4219f72d3ac698e475ae0f294dafeafaee248fd63887c"} Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.717657 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" event={"ID":"8b4af217-1e38-47e1-b306-a9d4d7ed90f8","Type":"ContainerStarted","Data":"a652e1a3e8ca41add08f46064de8dff5242ae2f7a8c4aa264e1136eb8241c37c"} Mar 20 16:12:55 crc kubenswrapper[4936]: I0320 16:12:55.723760 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-849fb747cf-nmlzn" event={"ID":"16f319af-481b-4398-84e2-a6ced6cbbd0c","Type":"ContainerStarted","Data":"5972b8789bec01e0bddb5f1dcac4cc1322a177ce423ff7715c81619d683e35e0"} Mar 20 16:12:56 crc kubenswrapper[4936]: E0320 16:12:56.067624 4936 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 20 16:12:56 crc kubenswrapper[4936]: E0320 16:12:56.067670 4936 configmap.go:193] Couldn't get configMap openshift-nmstate/nginx-conf: failed to sync configmap cache: timed out waiting for the condition Mar 20 16:12:56 crc kubenswrapper[4936]: E0320 16:12:56.067774 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00920150-af2f-46d4-a3cf-bdfea2640c89-plugin-serving-cert podName:00920150-af2f-46d4-a3cf-bdfea2640c89 nodeName:}" failed. No retries permitted until 2026-03-20 16:12:56.567745846 +0000 UTC m=+727.514113691 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/00920150-af2f-46d4-a3cf-bdfea2640c89-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-dk6hh" (UID: "00920150-af2f-46d4-a3cf-bdfea2640c89") : failed to sync secret cache: timed out waiting for the condition Mar 20 16:12:56 crc kubenswrapper[4936]: E0320 16:12:56.067807 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00920150-af2f-46d4-a3cf-bdfea2640c89-nginx-conf podName:00920150-af2f-46d4-a3cf-bdfea2640c89 nodeName:}" failed. No retries permitted until 2026-03-20 16:12:56.567791308 +0000 UTC m=+727.514159153 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/00920150-af2f-46d4-a3cf-bdfea2640c89-nginx-conf") pod "nmstate-console-plugin-86f58fcf4-dk6hh" (UID: "00920150-af2f-46d4-a3cf-bdfea2640c89") : failed to sync configmap cache: timed out waiting for the condition Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.103119 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.398431 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.494226 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-wrdsh" Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.590901 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/00920150-af2f-46d4-a3cf-bdfea2640c89-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.591053 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/00920150-af2f-46d4-a3cf-bdfea2640c89-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.592431 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/00920150-af2f-46d4-a3cf-bdfea2640c89-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.603649 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/00920150-af2f-46d4-a3cf-bdfea2640c89-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-dk6hh\" (UID: \"00920150-af2f-46d4-a3cf-bdfea2640c89\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.732015 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-849fb747cf-nmlzn" event={"ID":"16f319af-481b-4398-84e2-a6ced6cbbd0c","Type":"ContainerStarted","Data":"975385dc88feb0de2744f856a49dc7190e294f52de307ee2e33e4d9dafb11909"} Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.765302 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" Mar 20 16:12:56 crc kubenswrapper[4936]: I0320 16:12:56.774035 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-849fb747cf-nmlzn" podStartSLOduration=1.774016224 podStartE2EDuration="1.774016224s" podCreationTimestamp="2026-03-20 16:12:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:12:56.772639333 +0000 UTC m=+727.719007148" watchObservedRunningTime="2026-03-20 16:12:56.774016224 +0000 UTC m=+727.720384039" Mar 20 16:12:57 crc kubenswrapper[4936]: I0320 16:12:57.034696 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh"] Mar 20 16:12:57 crc kubenswrapper[4936]: I0320 16:12:57.741631 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" event={"ID":"00920150-af2f-46d4-a3cf-bdfea2640c89","Type":"ContainerStarted","Data":"185c76ceda66cf23c31330166651b856d64f1a713570fbab20a3405229b2c510"} Mar 20 16:12:58 crc kubenswrapper[4936]: I0320 16:12:58.231732 4936 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 16:12:58 crc kubenswrapper[4936]: I0320 16:12:58.758810 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" event={"ID":"8b4af217-1e38-47e1-b306-a9d4d7ed90f8","Type":"ContainerStarted","Data":"211408a59874dd369d19c69c9cb367a5c217e33175c99460383404da2b5fbef5"} Mar 20 16:12:58 crc kubenswrapper[4936]: I0320 16:12:58.759380 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:12:58 crc kubenswrapper[4936]: I0320 16:12:58.761414 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4" event={"ID":"f0f1b197-b717-44db-aed0-98ae53aba831","Type":"ContainerStarted","Data":"f3174607a8132f00e34a35f28128af9ca1a1c7d87027149303a968f8627ce54a"} Mar 20 16:12:58 crc kubenswrapper[4936]: I0320 16:12:58.855219 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:12:58 crc kubenswrapper[4936]: I0320 16:12:58.855276 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:12:58 crc kubenswrapper[4936]: I0320 16:12:58.855316 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:12:58 crc kubenswrapper[4936]: I0320 16:12:58.855827 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e12ccb52ca45cab384d2f9173c7abe5ae747d569fea13ff0f7cd5e5e7c20df0"} pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 16:12:58 crc kubenswrapper[4936]: I0320 16:12:58.855885 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" containerID="cri-o://6e12ccb52ca45cab384d2f9173c7abe5ae747d569fea13ff0f7cd5e5e7c20df0" gracePeriod=600 Mar 20 16:12:59 crc kubenswrapper[4936]: I0320 16:12:59.772159 4936 generic.go:334] "Generic (PLEG): container finished" podID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerID="6e12ccb52ca45cab384d2f9173c7abe5ae747d569fea13ff0f7cd5e5e7c20df0" exitCode=0 Mar 20 16:12:59 crc kubenswrapper[4936]: I0320 16:12:59.772243 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerDied","Data":"6e12ccb52ca45cab384d2f9173c7abe5ae747d569fea13ff0f7cd5e5e7c20df0"} Mar 20 16:12:59 crc kubenswrapper[4936]: I0320 16:12:59.772611 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"0421fad5e0fd4e5837ad1c312bd061ad68a03b4d2795d6099a681bc4b3ce8f6e"} Mar 20 16:12:59 crc kubenswrapper[4936]: I0320 16:12:59.772635 4936 scope.go:117] "RemoveContainer" containerID="ad7190bb15359c2cb63ff782fee447cfc84994fffc54cec84eef49413cfae9d9" Mar 20 16:12:59 crc kubenswrapper[4936]: I0320 16:12:59.775043 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-h59z7" event={"ID":"36c16ea1-7b8b-4175-8647-92fc61ee66d5","Type":"ContainerStarted","Data":"76a4a9ac5b27bd5f731b44e592d1e488a2447c63432c3a54071b486a5720e4c6"} Mar 20 16:12:59 crc kubenswrapper[4936]: I0320 16:12:59.775314 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:12:59 crc kubenswrapper[4936]: I0320 16:12:59.792816 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" podStartSLOduration=2.856548254 podStartE2EDuration="5.792796098s" podCreationTimestamp="2026-03-20 16:12:54 +0000 UTC" firstStartedPulling="2026-03-20 16:12:55.578774034 +0000 UTC m=+726.525141869" lastFinishedPulling="2026-03-20 16:12:58.515021888 +0000 UTC m=+729.461389713" observedRunningTime="2026-03-20 16:12:58.779584325 +0000 UTC m=+729.725952140" watchObservedRunningTime="2026-03-20 16:12:59.792796098 +0000 UTC m=+730.739163913" Mar 20 16:12:59 crc kubenswrapper[4936]: I0320 16:12:59.806069 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-h59z7" podStartSLOduration=2.520895064 podStartE2EDuration="5.806050388s" podCreationTimestamp="2026-03-20 16:12:54 +0000 UTC" firstStartedPulling="2026-03-20 16:12:55.188767535 +0000 UTC m=+726.135135350" lastFinishedPulling="2026-03-20 16:12:58.473922859 +0000 UTC m=+729.420290674" observedRunningTime="2026-03-20 16:12:59.801235427 +0000 UTC m=+730.747603242" watchObservedRunningTime="2026-03-20 16:12:59.806050388 +0000 UTC m=+730.752418223" Mar 20 16:13:01 crc kubenswrapper[4936]: I0320 16:13:01.791974 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" event={"ID":"00920150-af2f-46d4-a3cf-bdfea2640c89","Type":"ContainerStarted","Data":"37b05e1ff7236a9d1e1b876b03cf9abde9321607c44296c0920d2e93ca0496d8"} Mar 20 16:13:01 crc kubenswrapper[4936]: I0320 16:13:01.818399 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-dk6hh" podStartSLOduration=3.608911147 podStartE2EDuration="7.818360277s" podCreationTimestamp="2026-03-20 16:12:54 +0000 UTC" firstStartedPulling="2026-03-20 16:12:57.045383061 +0000 UTC m=+727.991750886" lastFinishedPulling="2026-03-20 16:13:01.254832201 +0000 UTC m=+732.201200016" observedRunningTime="2026-03-20 16:13:01.809504027 +0000 UTC m=+732.755871842" watchObservedRunningTime="2026-03-20 16:13:01.818360277 +0000 UTC m=+732.764728132" Mar 20 16:13:02 crc kubenswrapper[4936]: I0320 16:13:02.805036 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4" event={"ID":"f0f1b197-b717-44db-aed0-98ae53aba831","Type":"ContainerStarted","Data":"4381aa7da287074d92734593c323c0ac7bbcafd70c5ab97c35930700f13978b4"} Mar 20 16:13:02 crc kubenswrapper[4936]: I0320 16:13:02.840225 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-mmtw4" podStartSLOduration=2.521447851 podStartE2EDuration="8.840191544s" podCreationTimestamp="2026-03-20 16:12:54 +0000 UTC" firstStartedPulling="2026-03-20 16:12:55.646996702 +0000 UTC m=+726.593364537" lastFinishedPulling="2026-03-20 16:13:01.965740425 +0000 UTC m=+732.912108230" observedRunningTime="2026-03-20 16:13:02.829305913 +0000 UTC m=+733.775673808" watchObservedRunningTime="2026-03-20 16:13:02.840191544 +0000 UTC m=+733.786559399" Mar 20 16:13:05 crc kubenswrapper[4936]: I0320 16:13:05.183506 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-h59z7" Mar 20 16:13:05 crc kubenswrapper[4936]: I0320 16:13:05.496437 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:13:05 crc kubenswrapper[4936]: I0320 16:13:05.496522 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:13:05 crc kubenswrapper[4936]: I0320 16:13:05.501883 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:13:05 crc kubenswrapper[4936]: I0320 16:13:05.829851 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-849fb747cf-nmlzn" Mar 20 16:13:05 crc kubenswrapper[4936]: I0320 16:13:05.889949 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6q5fl"] Mar 20 16:13:15 crc kubenswrapper[4936]: I0320 16:13:15.137102 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-dqbtr" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.683567 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj"] Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.686082 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.689313 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.694887 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj"] Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.798648 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.798700 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.798858 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82vqt\" (UniqueName: \"kubernetes.io/projected/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-kube-api-access-82vqt\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.900111 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.900451 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.900511 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82vqt\" (UniqueName: \"kubernetes.io/projected/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-kube-api-access-82vqt\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.901122 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.901391 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:29 crc kubenswrapper[4936]: I0320 16:13:29.928280 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82vqt\" (UniqueName: \"kubernetes.io/projected/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-kube-api-access-82vqt\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:30 crc kubenswrapper[4936]: I0320 16:13:30.007708 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:30 crc kubenswrapper[4936]: I0320 16:13:30.438925 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj"] Mar 20 16:13:30 crc kubenswrapper[4936]: I0320 16:13:30.938647 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-6q5fl" podUID="c52c0329-ca68-4d95-8341-ab8191c3a186" containerName="console" containerID="cri-o://6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0" gracePeriod=15 Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.002271 4936 generic.go:334] "Generic (PLEG): container finished" podID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerID="65d4ddf7eb76af4da4fce158821c948c1230b755056bb72773fd4804fa3b350a" exitCode=0 Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.002383 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" event={"ID":"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10","Type":"ContainerDied","Data":"65d4ddf7eb76af4da4fce158821c948c1230b755056bb72773fd4804fa3b350a"} Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.003279 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" event={"ID":"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10","Type":"ContainerStarted","Data":"237389ddc2ce4a4d43c3da764455285ca2ae6858c79a9e1204f0421fd91aaddf"} Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.005071 4936 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.381055 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6q5fl_c52c0329-ca68-4d95-8341-ab8191c3a186/console/0.log" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.381422 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.441729 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-trusted-ca-bundle\") pod \"c52c0329-ca68-4d95-8341-ab8191c3a186\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.441837 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-oauth-config\") pod \"c52c0329-ca68-4d95-8341-ab8191c3a186\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.441968 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-oauth-serving-cert\") pod \"c52c0329-ca68-4d95-8341-ab8191c3a186\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.442059 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpxkx\" (UniqueName: \"kubernetes.io/projected/c52c0329-ca68-4d95-8341-ab8191c3a186-kube-api-access-wpxkx\") pod \"c52c0329-ca68-4d95-8341-ab8191c3a186\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.442529 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-console-config\") pod \"c52c0329-ca68-4d95-8341-ab8191c3a186\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.442579 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-serving-cert\") pod \"c52c0329-ca68-4d95-8341-ab8191c3a186\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.442899 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c52c0329-ca68-4d95-8341-ab8191c3a186" (UID: "c52c0329-ca68-4d95-8341-ab8191c3a186"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.442610 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-service-ca\") pod \"c52c0329-ca68-4d95-8341-ab8191c3a186\" (UID: \"c52c0329-ca68-4d95-8341-ab8191c3a186\") " Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.443333 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c52c0329-ca68-4d95-8341-ab8191c3a186" (UID: "c52c0329-ca68-4d95-8341-ab8191c3a186"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.443368 4936 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.443440 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-service-ca" (OuterVolumeSpecName: "service-ca") pod "c52c0329-ca68-4d95-8341-ab8191c3a186" (UID: "c52c0329-ca68-4d95-8341-ab8191c3a186"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.443898 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-console-config" (OuterVolumeSpecName: "console-config") pod "c52c0329-ca68-4d95-8341-ab8191c3a186" (UID: "c52c0329-ca68-4d95-8341-ab8191c3a186"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.448413 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c52c0329-ca68-4d95-8341-ab8191c3a186" (UID: "c52c0329-ca68-4d95-8341-ab8191c3a186"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.449645 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c52c0329-ca68-4d95-8341-ab8191c3a186-kube-api-access-wpxkx" (OuterVolumeSpecName: "kube-api-access-wpxkx") pod "c52c0329-ca68-4d95-8341-ab8191c3a186" (UID: "c52c0329-ca68-4d95-8341-ab8191c3a186"). InnerVolumeSpecName "kube-api-access-wpxkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.450152 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c52c0329-ca68-4d95-8341-ab8191c3a186" (UID: "c52c0329-ca68-4d95-8341-ab8191c3a186"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.544564 4936 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.544664 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpxkx\" (UniqueName: \"kubernetes.io/projected/c52c0329-ca68-4d95-8341-ab8191c3a186-kube-api-access-wpxkx\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.544681 4936 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-console-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.544691 4936 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c52c0329-ca68-4d95-8341-ab8191c3a186-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.544702 4936 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:31 crc kubenswrapper[4936]: I0320 16:13:31.544712 4936 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c52c0329-ca68-4d95-8341-ab8191c3a186-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.012860 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6q5fl_c52c0329-ca68-4d95-8341-ab8191c3a186/console/0.log" Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.012957 4936 generic.go:334] "Generic (PLEG): container finished" podID="c52c0329-ca68-4d95-8341-ab8191c3a186" containerID="6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0" exitCode=2 Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.013008 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6q5fl" event={"ID":"c52c0329-ca68-4d95-8341-ab8191c3a186","Type":"ContainerDied","Data":"6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0"} Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.013047 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6q5fl" Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.013063 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6q5fl" event={"ID":"c52c0329-ca68-4d95-8341-ab8191c3a186","Type":"ContainerDied","Data":"225ff718e9e7e6534de7e30034b9c3c82fa1cb352430832486bc8078bd1791b4"} Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.013094 4936 scope.go:117] "RemoveContainer" containerID="6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0" Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.055391 4936 scope.go:117] "RemoveContainer" containerID="6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0" Mar 20 16:13:32 crc kubenswrapper[4936]: E0320 16:13:32.056692 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0\": container with ID starting with 6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0 not found: ID does not exist" containerID="6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0" Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.056745 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0"} err="failed to get container status \"6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0\": rpc error: code = NotFound desc = could not find container \"6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0\": container with ID starting with 6948bdab6bda5be6e9dece0f6c3c08d9c1a0ef719ebbaf7d3596fae3845fbba0 not found: ID does not exist" Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.058094 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6q5fl"] Mar 20 16:13:32 crc kubenswrapper[4936]: I0320 16:13:32.063530 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-6q5fl"] Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.022221 4936 generic.go:334] "Generic (PLEG): container finished" podID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerID="c4b649b6a59a07e5d86bbb5cc35b8b1e01d4efaf004beb1fb35e88b06335c71d" exitCode=0 Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.022310 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" event={"ID":"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10","Type":"ContainerDied","Data":"c4b649b6a59a07e5d86bbb5cc35b8b1e01d4efaf004beb1fb35e88b06335c71d"} Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.047447 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w5fd2"] Mar 20 16:13:33 crc kubenswrapper[4936]: E0320 16:13:33.047770 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c52c0329-ca68-4d95-8341-ab8191c3a186" containerName="console" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.047796 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="c52c0329-ca68-4d95-8341-ab8191c3a186" containerName="console" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.048017 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="c52c0329-ca68-4d95-8341-ab8191c3a186" containerName="console" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.049304 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.058844 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5fd2"] Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.170726 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-utilities\") pod \"redhat-operators-w5fd2\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.170808 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp2p9\" (UniqueName: \"kubernetes.io/projected/7366f736-aa9a-4a46-a07c-4a45ee6c3368-kube-api-access-gp2p9\") pod \"redhat-operators-w5fd2\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.170872 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-catalog-content\") pod \"redhat-operators-w5fd2\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.271832 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-utilities\") pod \"redhat-operators-w5fd2\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.271897 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp2p9\" (UniqueName: \"kubernetes.io/projected/7366f736-aa9a-4a46-a07c-4a45ee6c3368-kube-api-access-gp2p9\") pod \"redhat-operators-w5fd2\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.271937 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-catalog-content\") pod \"redhat-operators-w5fd2\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.272407 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-utilities\") pod \"redhat-operators-w5fd2\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.272453 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-catalog-content\") pod \"redhat-operators-w5fd2\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.293716 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp2p9\" (UniqueName: \"kubernetes.io/projected/7366f736-aa9a-4a46-a07c-4a45ee6c3368-kube-api-access-gp2p9\") pod \"redhat-operators-w5fd2\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.379893 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.606920 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5fd2"] Mar 20 16:13:33 crc kubenswrapper[4936]: W0320 16:13:33.614681 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7366f736_aa9a_4a46_a07c_4a45ee6c3368.slice/crio-7e37b77878625da744fe9452d8badb01a684bbb392b954fbbc33d50a7530188f WatchSource:0}: Error finding container 7e37b77878625da744fe9452d8badb01a684bbb392b954fbbc33d50a7530188f: Status 404 returned error can't find the container with id 7e37b77878625da744fe9452d8badb01a684bbb392b954fbbc33d50a7530188f Mar 20 16:13:33 crc kubenswrapper[4936]: I0320 16:13:33.863138 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c52c0329-ca68-4d95-8341-ab8191c3a186" path="/var/lib/kubelet/pods/c52c0329-ca68-4d95-8341-ab8191c3a186/volumes" Mar 20 16:13:34 crc kubenswrapper[4936]: I0320 16:13:34.030961 4936 generic.go:334] "Generic (PLEG): container finished" podID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerID="991177fa8872015e71a2051c562fa299be32cf9d716a08a8e8feae3ee6e3113b" exitCode=0 Mar 20 16:13:34 crc kubenswrapper[4936]: I0320 16:13:34.031030 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" event={"ID":"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10","Type":"ContainerDied","Data":"991177fa8872015e71a2051c562fa299be32cf9d716a08a8e8feae3ee6e3113b"} Mar 20 16:13:34 crc kubenswrapper[4936]: I0320 16:13:34.032700 4936 generic.go:334] "Generic (PLEG): container finished" podID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerID="d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb" exitCode=0 Mar 20 16:13:34 crc kubenswrapper[4936]: I0320 16:13:34.032729 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5fd2" event={"ID":"7366f736-aa9a-4a46-a07c-4a45ee6c3368","Type":"ContainerDied","Data":"d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb"} Mar 20 16:13:34 crc kubenswrapper[4936]: I0320 16:13:34.032755 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5fd2" event={"ID":"7366f736-aa9a-4a46-a07c-4a45ee6c3368","Type":"ContainerStarted","Data":"7e37b77878625da744fe9452d8badb01a684bbb392b954fbbc33d50a7530188f"} Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.329084 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.402742 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82vqt\" (UniqueName: \"kubernetes.io/projected/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-kube-api-access-82vqt\") pod \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.402808 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-util\") pod \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.402877 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-bundle\") pod \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\" (UID: \"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10\") " Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.403853 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-bundle" (OuterVolumeSpecName: "bundle") pod "ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" (UID: "ece6daad-afcb-4bfb-ace9-1fb25a6c9f10"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.407996 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-kube-api-access-82vqt" (OuterVolumeSpecName: "kube-api-access-82vqt") pod "ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" (UID: "ece6daad-afcb-4bfb-ace9-1fb25a6c9f10"). InnerVolumeSpecName "kube-api-access-82vqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.416257 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-util" (OuterVolumeSpecName: "util") pod "ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" (UID: "ece6daad-afcb-4bfb-ace9-1fb25a6c9f10"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.504199 4936 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.504237 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82vqt\" (UniqueName: \"kubernetes.io/projected/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-kube-api-access-82vqt\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:35 crc kubenswrapper[4936]: I0320 16:13:35.504252 4936 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ece6daad-afcb-4bfb-ace9-1fb25a6c9f10-util\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:36 crc kubenswrapper[4936]: I0320 16:13:36.046639 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" event={"ID":"ece6daad-afcb-4bfb-ace9-1fb25a6c9f10","Type":"ContainerDied","Data":"237389ddc2ce4a4d43c3da764455285ca2ae6858c79a9e1204f0421fd91aaddf"} Mar 20 16:13:36 crc kubenswrapper[4936]: I0320 16:13:36.046677 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="237389ddc2ce4a4d43c3da764455285ca2ae6858c79a9e1204f0421fd91aaddf" Mar 20 16:13:36 crc kubenswrapper[4936]: I0320 16:13:36.046738 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj" Mar 20 16:13:36 crc kubenswrapper[4936]: I0320 16:13:36.050456 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5fd2" event={"ID":"7366f736-aa9a-4a46-a07c-4a45ee6c3368","Type":"ContainerStarted","Data":"86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086"} Mar 20 16:13:37 crc kubenswrapper[4936]: I0320 16:13:37.062876 4936 generic.go:334] "Generic (PLEG): container finished" podID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerID="86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086" exitCode=0 Mar 20 16:13:37 crc kubenswrapper[4936]: I0320 16:13:37.062920 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5fd2" event={"ID":"7366f736-aa9a-4a46-a07c-4a45ee6c3368","Type":"ContainerDied","Data":"86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086"} Mar 20 16:13:38 crc kubenswrapper[4936]: I0320 16:13:38.070320 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5fd2" event={"ID":"7366f736-aa9a-4a46-a07c-4a45ee6c3368","Type":"ContainerStarted","Data":"6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7"} Mar 20 16:13:38 crc kubenswrapper[4936]: I0320 16:13:38.090444 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w5fd2" podStartSLOduration=1.624796407 podStartE2EDuration="5.090429073s" podCreationTimestamp="2026-03-20 16:13:33 +0000 UTC" firstStartedPulling="2026-03-20 16:13:34.034233974 +0000 UTC m=+764.980601799" lastFinishedPulling="2026-03-20 16:13:37.49986661 +0000 UTC m=+768.446234465" observedRunningTime="2026-03-20 16:13:38.087604589 +0000 UTC m=+769.033972424" watchObservedRunningTime="2026-03-20 16:13:38.090429073 +0000 UTC m=+769.036796888" Mar 20 16:13:43 crc kubenswrapper[4936]: I0320 16:13:43.380487 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:43 crc kubenswrapper[4936]: I0320 16:13:43.382312 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:44 crc kubenswrapper[4936]: I0320 16:13:44.468393 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w5fd2" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerName="registry-server" probeResult="failure" output=< Mar 20 16:13:44 crc kubenswrapper[4936]: timeout: failed to connect service ":50051" within 1s Mar 20 16:13:44 crc kubenswrapper[4936]: > Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.697709 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7"] Mar 20 16:13:45 crc kubenswrapper[4936]: E0320 16:13:45.698289 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerName="util" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.698306 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerName="util" Mar 20 16:13:45 crc kubenswrapper[4936]: E0320 16:13:45.698328 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerName="extract" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.698336 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerName="extract" Mar 20 16:13:45 crc kubenswrapper[4936]: E0320 16:13:45.698349 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerName="pull" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.698356 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerName="pull" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.698474 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ece6daad-afcb-4bfb-ace9-1fb25a6c9f10" containerName="extract" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.698939 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.700364 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.700828 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.701610 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-dbfmx" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.702180 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.703574 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.717520 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7"] Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.835696 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b880eadb-7994-485a-99fc-f7a4265707d3-apiservice-cert\") pod \"metallb-operator-controller-manager-7db57c7955-sxrb7\" (UID: \"b880eadb-7994-485a-99fc-f7a4265707d3\") " pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.835777 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b880eadb-7994-485a-99fc-f7a4265707d3-webhook-cert\") pod \"metallb-operator-controller-manager-7db57c7955-sxrb7\" (UID: \"b880eadb-7994-485a-99fc-f7a4265707d3\") " pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.835810 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7zmg\" (UniqueName: \"kubernetes.io/projected/b880eadb-7994-485a-99fc-f7a4265707d3-kube-api-access-n7zmg\") pod \"metallb-operator-controller-manager-7db57c7955-sxrb7\" (UID: \"b880eadb-7994-485a-99fc-f7a4265707d3\") " pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.937410 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b880eadb-7994-485a-99fc-f7a4265707d3-apiservice-cert\") pod \"metallb-operator-controller-manager-7db57c7955-sxrb7\" (UID: \"b880eadb-7994-485a-99fc-f7a4265707d3\") " pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.937514 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b880eadb-7994-485a-99fc-f7a4265707d3-webhook-cert\") pod \"metallb-operator-controller-manager-7db57c7955-sxrb7\" (UID: \"b880eadb-7994-485a-99fc-f7a4265707d3\") " pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.937559 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7zmg\" (UniqueName: \"kubernetes.io/projected/b880eadb-7994-485a-99fc-f7a4265707d3-kube-api-access-n7zmg\") pod \"metallb-operator-controller-manager-7db57c7955-sxrb7\" (UID: \"b880eadb-7994-485a-99fc-f7a4265707d3\") " pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.943204 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz"] Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.944024 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b880eadb-7994-485a-99fc-f7a4265707d3-apiservice-cert\") pod \"metallb-operator-controller-manager-7db57c7955-sxrb7\" (UID: \"b880eadb-7994-485a-99fc-f7a4265707d3\") " pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.944611 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.946392 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b880eadb-7994-485a-99fc-f7a4265707d3-webhook-cert\") pod \"metallb-operator-controller-manager-7db57c7955-sxrb7\" (UID: \"b880eadb-7994-485a-99fc-f7a4265707d3\") " pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.948041 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.948519 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.959418 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-cdh6f" Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.967435 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz"] Mar 20 16:13:45 crc kubenswrapper[4936]: I0320 16:13:45.986531 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7zmg\" (UniqueName: \"kubernetes.io/projected/b880eadb-7994-485a-99fc-f7a4265707d3-kube-api-access-n7zmg\") pod \"metallb-operator-controller-manager-7db57c7955-sxrb7\" (UID: \"b880eadb-7994-485a-99fc-f7a4265707d3\") " pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.013152 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.044231 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k2bz\" (UniqueName: \"kubernetes.io/projected/05880ff9-4310-40a0-a8f3-47dfc79f87e2-kube-api-access-9k2bz\") pod \"metallb-operator-webhook-server-5b8d4cf8dd-js5nz\" (UID: \"05880ff9-4310-40a0-a8f3-47dfc79f87e2\") " pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.044627 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/05880ff9-4310-40a0-a8f3-47dfc79f87e2-webhook-cert\") pod \"metallb-operator-webhook-server-5b8d4cf8dd-js5nz\" (UID: \"05880ff9-4310-40a0-a8f3-47dfc79f87e2\") " pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.044691 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/05880ff9-4310-40a0-a8f3-47dfc79f87e2-apiservice-cert\") pod \"metallb-operator-webhook-server-5b8d4cf8dd-js5nz\" (UID: \"05880ff9-4310-40a0-a8f3-47dfc79f87e2\") " pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.148093 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k2bz\" (UniqueName: \"kubernetes.io/projected/05880ff9-4310-40a0-a8f3-47dfc79f87e2-kube-api-access-9k2bz\") pod \"metallb-operator-webhook-server-5b8d4cf8dd-js5nz\" (UID: \"05880ff9-4310-40a0-a8f3-47dfc79f87e2\") " pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.148140 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/05880ff9-4310-40a0-a8f3-47dfc79f87e2-webhook-cert\") pod \"metallb-operator-webhook-server-5b8d4cf8dd-js5nz\" (UID: \"05880ff9-4310-40a0-a8f3-47dfc79f87e2\") " pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.148174 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/05880ff9-4310-40a0-a8f3-47dfc79f87e2-apiservice-cert\") pod \"metallb-operator-webhook-server-5b8d4cf8dd-js5nz\" (UID: \"05880ff9-4310-40a0-a8f3-47dfc79f87e2\") " pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.156328 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/05880ff9-4310-40a0-a8f3-47dfc79f87e2-apiservice-cert\") pod \"metallb-operator-webhook-server-5b8d4cf8dd-js5nz\" (UID: \"05880ff9-4310-40a0-a8f3-47dfc79f87e2\") " pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.173796 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/05880ff9-4310-40a0-a8f3-47dfc79f87e2-webhook-cert\") pod \"metallb-operator-webhook-server-5b8d4cf8dd-js5nz\" (UID: \"05880ff9-4310-40a0-a8f3-47dfc79f87e2\") " pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.179235 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k2bz\" (UniqueName: \"kubernetes.io/projected/05880ff9-4310-40a0-a8f3-47dfc79f87e2-kube-api-access-9k2bz\") pod \"metallb-operator-webhook-server-5b8d4cf8dd-js5nz\" (UID: \"05880ff9-4310-40a0-a8f3-47dfc79f87e2\") " pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.325907 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.494371 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7"] Mar 20 16:13:46 crc kubenswrapper[4936]: W0320 16:13:46.505300 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb880eadb_7994_485a_99fc_f7a4265707d3.slice/crio-2f2b39fbf29bfdb3706bf9b7203f27c7cd3f1e9a640151eca1ff041bb7ea70b8 WatchSource:0}: Error finding container 2f2b39fbf29bfdb3706bf9b7203f27c7cd3f1e9a640151eca1ff041bb7ea70b8: Status 404 returned error can't find the container with id 2f2b39fbf29bfdb3706bf9b7203f27c7cd3f1e9a640151eca1ff041bb7ea70b8 Mar 20 16:13:46 crc kubenswrapper[4936]: I0320 16:13:46.525276 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz"] Mar 20 16:13:46 crc kubenswrapper[4936]: W0320 16:13:46.542069 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05880ff9_4310_40a0_a8f3_47dfc79f87e2.slice/crio-4ebb0a168af63c78f15c119b0d6c8ba2448cd67ae42acc46801872f4d0b2dc69 WatchSource:0}: Error finding container 4ebb0a168af63c78f15c119b0d6c8ba2448cd67ae42acc46801872f4d0b2dc69: Status 404 returned error can't find the container with id 4ebb0a168af63c78f15c119b0d6c8ba2448cd67ae42acc46801872f4d0b2dc69 Mar 20 16:13:47 crc kubenswrapper[4936]: I0320 16:13:47.120495 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" event={"ID":"05880ff9-4310-40a0-a8f3-47dfc79f87e2","Type":"ContainerStarted","Data":"4ebb0a168af63c78f15c119b0d6c8ba2448cd67ae42acc46801872f4d0b2dc69"} Mar 20 16:13:47 crc kubenswrapper[4936]: I0320 16:13:47.121961 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" event={"ID":"b880eadb-7994-485a-99fc-f7a4265707d3","Type":"ContainerStarted","Data":"2f2b39fbf29bfdb3706bf9b7203f27c7cd3f1e9a640151eca1ff041bb7ea70b8"} Mar 20 16:13:51 crc kubenswrapper[4936]: I0320 16:13:51.150901 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" event={"ID":"b880eadb-7994-485a-99fc-f7a4265707d3","Type":"ContainerStarted","Data":"028a3f2198081be84b3878b6975532fecf960d2c3344436c96d5dfa35c641468"} Mar 20 16:13:51 crc kubenswrapper[4936]: I0320 16:13:51.151685 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:13:51 crc kubenswrapper[4936]: I0320 16:13:51.180120 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" podStartSLOduration=2.23796866 podStartE2EDuration="6.180092258s" podCreationTimestamp="2026-03-20 16:13:45 +0000 UTC" firstStartedPulling="2026-03-20 16:13:46.507944 +0000 UTC m=+777.454311815" lastFinishedPulling="2026-03-20 16:13:50.450067598 +0000 UTC m=+781.396435413" observedRunningTime="2026-03-20 16:13:51.177637546 +0000 UTC m=+782.124005401" watchObservedRunningTime="2026-03-20 16:13:51.180092258 +0000 UTC m=+782.126460123" Mar 20 16:13:53 crc kubenswrapper[4936]: I0320 16:13:53.429387 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:53 crc kubenswrapper[4936]: I0320 16:13:53.474634 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:53 crc kubenswrapper[4936]: I0320 16:13:53.842618 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5fd2"] Mar 20 16:13:54 crc kubenswrapper[4936]: I0320 16:13:54.174161 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" event={"ID":"05880ff9-4310-40a0-a8f3-47dfc79f87e2","Type":"ContainerStarted","Data":"0ce71d80518539c2db15b5cc2a1c9e1aba29f528cfbcd20feef4e91519765cfe"} Mar 20 16:13:54 crc kubenswrapper[4936]: I0320 16:13:54.174231 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:13:54 crc kubenswrapper[4936]: I0320 16:13:54.194050 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" podStartSLOduration=2.536783007 podStartE2EDuration="9.194030112s" podCreationTimestamp="2026-03-20 16:13:45 +0000 UTC" firstStartedPulling="2026-03-20 16:13:46.546015441 +0000 UTC m=+777.492383256" lastFinishedPulling="2026-03-20 16:13:53.203262546 +0000 UTC m=+784.149630361" observedRunningTime="2026-03-20 16:13:54.19360823 +0000 UTC m=+785.139976065" watchObservedRunningTime="2026-03-20 16:13:54.194030112 +0000 UTC m=+785.140397927" Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.179610 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w5fd2" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerName="registry-server" containerID="cri-o://6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7" gracePeriod=2 Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.554324 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.696852 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-utilities\") pod \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.697227 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gp2p9\" (UniqueName: \"kubernetes.io/projected/7366f736-aa9a-4a46-a07c-4a45ee6c3368-kube-api-access-gp2p9\") pod \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.697294 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-catalog-content\") pod \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\" (UID: \"7366f736-aa9a-4a46-a07c-4a45ee6c3368\") " Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.697662 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-utilities" (OuterVolumeSpecName: "utilities") pod "7366f736-aa9a-4a46-a07c-4a45ee6c3368" (UID: "7366f736-aa9a-4a46-a07c-4a45ee6c3368"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.705441 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7366f736-aa9a-4a46-a07c-4a45ee6c3368-kube-api-access-gp2p9" (OuterVolumeSpecName: "kube-api-access-gp2p9") pod "7366f736-aa9a-4a46-a07c-4a45ee6c3368" (UID: "7366f736-aa9a-4a46-a07c-4a45ee6c3368"). InnerVolumeSpecName "kube-api-access-gp2p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.799112 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gp2p9\" (UniqueName: \"kubernetes.io/projected/7366f736-aa9a-4a46-a07c-4a45ee6c3368-kube-api-access-gp2p9\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.799156 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.850647 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7366f736-aa9a-4a46-a07c-4a45ee6c3368" (UID: "7366f736-aa9a-4a46-a07c-4a45ee6c3368"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:13:55 crc kubenswrapper[4936]: I0320 16:13:55.900079 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7366f736-aa9a-4a46-a07c-4a45ee6c3368-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.188163 4936 generic.go:334] "Generic (PLEG): container finished" podID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerID="6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7" exitCode=0 Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.188204 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5fd2" event={"ID":"7366f736-aa9a-4a46-a07c-4a45ee6c3368","Type":"ContainerDied","Data":"6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7"} Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.188229 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5fd2" event={"ID":"7366f736-aa9a-4a46-a07c-4a45ee6c3368","Type":"ContainerDied","Data":"7e37b77878625da744fe9452d8badb01a684bbb392b954fbbc33d50a7530188f"} Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.188245 4936 scope.go:117] "RemoveContainer" containerID="6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.188275 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5fd2" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.209100 4936 scope.go:117] "RemoveContainer" containerID="86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.215935 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5fd2"] Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.242851 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w5fd2"] Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.244989 4936 scope.go:117] "RemoveContainer" containerID="d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.274650 4936 scope.go:117] "RemoveContainer" containerID="6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7" Mar 20 16:13:56 crc kubenswrapper[4936]: E0320 16:13:56.275299 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7\": container with ID starting with 6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7 not found: ID does not exist" containerID="6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.275337 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7"} err="failed to get container status \"6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7\": rpc error: code = NotFound desc = could not find container \"6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7\": container with ID starting with 6ecbe6381e2b7e814465c37b26e24a78c83c4c3b344ee3d05c0f25b8cb9b5bc7 not found: ID does not exist" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.275383 4936 scope.go:117] "RemoveContainer" containerID="86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086" Mar 20 16:13:56 crc kubenswrapper[4936]: E0320 16:13:56.276016 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086\": container with ID starting with 86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086 not found: ID does not exist" containerID="86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.276060 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086"} err="failed to get container status \"86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086\": rpc error: code = NotFound desc = could not find container \"86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086\": container with ID starting with 86e89abb8da13e9aa7e8084c3cbfe99d9899cfd321af885636dbdc20fcd6e086 not found: ID does not exist" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.276094 4936 scope.go:117] "RemoveContainer" containerID="d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb" Mar 20 16:13:56 crc kubenswrapper[4936]: E0320 16:13:56.276571 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb\": container with ID starting with d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb not found: ID does not exist" containerID="d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb" Mar 20 16:13:56 crc kubenswrapper[4936]: I0320 16:13:56.276598 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb"} err="failed to get container status \"d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb\": rpc error: code = NotFound desc = could not find container \"d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb\": container with ID starting with d8ae39af8e0d219c35254585cdf1485f33117e42a2c03b58b4104c9073b149bb not found: ID does not exist" Mar 20 16:13:57 crc kubenswrapper[4936]: I0320 16:13:57.885292 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" path="/var/lib/kubelet/pods/7366f736-aa9a-4a46-a07c-4a45ee6c3368/volumes" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.124589 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567054-xkcjv"] Mar 20 16:14:00 crc kubenswrapper[4936]: E0320 16:14:00.124861 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerName="registry-server" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.124877 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerName="registry-server" Mar 20 16:14:00 crc kubenswrapper[4936]: E0320 16:14:00.124893 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerName="extract-utilities" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.124901 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerName="extract-utilities" Mar 20 16:14:00 crc kubenswrapper[4936]: E0320 16:14:00.124918 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerName="extract-content" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.124926 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerName="extract-content" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.125045 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="7366f736-aa9a-4a46-a07c-4a45ee6c3368" containerName="registry-server" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.125459 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567054-xkcjv" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.127373 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.134943 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.135867 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.139089 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567054-xkcjv"] Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.239794 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4n9l\" (UniqueName: \"kubernetes.io/projected/89ff8a26-ef18-4434-9601-11a47af62d20-kube-api-access-m4n9l\") pod \"auto-csr-approver-29567054-xkcjv\" (UID: \"89ff8a26-ef18-4434-9601-11a47af62d20\") " pod="openshift-infra/auto-csr-approver-29567054-xkcjv" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.340707 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4n9l\" (UniqueName: \"kubernetes.io/projected/89ff8a26-ef18-4434-9601-11a47af62d20-kube-api-access-m4n9l\") pod \"auto-csr-approver-29567054-xkcjv\" (UID: \"89ff8a26-ef18-4434-9601-11a47af62d20\") " pod="openshift-infra/auto-csr-approver-29567054-xkcjv" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.358123 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4n9l\" (UniqueName: \"kubernetes.io/projected/89ff8a26-ef18-4434-9601-11a47af62d20-kube-api-access-m4n9l\") pod \"auto-csr-approver-29567054-xkcjv\" (UID: \"89ff8a26-ef18-4434-9601-11a47af62d20\") " pod="openshift-infra/auto-csr-approver-29567054-xkcjv" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.443984 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567054-xkcjv" Mar 20 16:14:00 crc kubenswrapper[4936]: I0320 16:14:00.841729 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567054-xkcjv"] Mar 20 16:14:01 crc kubenswrapper[4936]: I0320 16:14:01.220376 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567054-xkcjv" event={"ID":"89ff8a26-ef18-4434-9601-11a47af62d20","Type":"ContainerStarted","Data":"af5a7ace30cefd71e74ead25b08c63f0d10332d8b8ed1b8cbda6b7bb51ba80ec"} Mar 20 16:14:06 crc kubenswrapper[4936]: I0320 16:14:06.331606 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5b8d4cf8dd-js5nz" Mar 20 16:14:07 crc kubenswrapper[4936]: I0320 16:14:07.262258 4936 generic.go:334] "Generic (PLEG): container finished" podID="89ff8a26-ef18-4434-9601-11a47af62d20" containerID="a37b6de81e8afb023b8879d76ef9385e8c34b13ed4f56cd1c86214ce40d0c5f3" exitCode=0 Mar 20 16:14:07 crc kubenswrapper[4936]: I0320 16:14:07.262339 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567054-xkcjv" event={"ID":"89ff8a26-ef18-4434-9601-11a47af62d20","Type":"ContainerDied","Data":"a37b6de81e8afb023b8879d76ef9385e8c34b13ed4f56cd1c86214ce40d0c5f3"} Mar 20 16:14:08 crc kubenswrapper[4936]: I0320 16:14:08.557262 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567054-xkcjv" Mar 20 16:14:08 crc kubenswrapper[4936]: I0320 16:14:08.649266 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4n9l\" (UniqueName: \"kubernetes.io/projected/89ff8a26-ef18-4434-9601-11a47af62d20-kube-api-access-m4n9l\") pod \"89ff8a26-ef18-4434-9601-11a47af62d20\" (UID: \"89ff8a26-ef18-4434-9601-11a47af62d20\") " Mar 20 16:14:08 crc kubenswrapper[4936]: I0320 16:14:08.657279 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89ff8a26-ef18-4434-9601-11a47af62d20-kube-api-access-m4n9l" (OuterVolumeSpecName: "kube-api-access-m4n9l") pod "89ff8a26-ef18-4434-9601-11a47af62d20" (UID: "89ff8a26-ef18-4434-9601-11a47af62d20"). InnerVolumeSpecName "kube-api-access-m4n9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:14:08 crc kubenswrapper[4936]: I0320 16:14:08.750783 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4n9l\" (UniqueName: \"kubernetes.io/projected/89ff8a26-ef18-4434-9601-11a47af62d20-kube-api-access-m4n9l\") on node \"crc\" DevicePath \"\"" Mar 20 16:14:09 crc kubenswrapper[4936]: I0320 16:14:09.280292 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567054-xkcjv" event={"ID":"89ff8a26-ef18-4434-9601-11a47af62d20","Type":"ContainerDied","Data":"af5a7ace30cefd71e74ead25b08c63f0d10332d8b8ed1b8cbda6b7bb51ba80ec"} Mar 20 16:14:09 crc kubenswrapper[4936]: I0320 16:14:09.280333 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af5a7ace30cefd71e74ead25b08c63f0d10332d8b8ed1b8cbda6b7bb51ba80ec" Mar 20 16:14:09 crc kubenswrapper[4936]: I0320 16:14:09.280390 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567054-xkcjv" Mar 20 16:14:09 crc kubenswrapper[4936]: I0320 16:14:09.619037 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567048-c4c4x"] Mar 20 16:14:09 crc kubenswrapper[4936]: I0320 16:14:09.626750 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567048-c4c4x"] Mar 20 16:14:09 crc kubenswrapper[4936]: I0320 16:14:09.866506 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e10e6f00-e11c-4f35-abad-7127db0c7cf3" path="/var/lib/kubelet/pods/e10e6f00-e11c-4f35-abad-7127db0c7cf3/volumes" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.016893 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7db57c7955-sxrb7" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.721123 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-4fg6x"] Mar 20 16:14:26 crc kubenswrapper[4936]: E0320 16:14:26.721391 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89ff8a26-ef18-4434-9601-11a47af62d20" containerName="oc" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.721412 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="89ff8a26-ef18-4434-9601-11a47af62d20" containerName="oc" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.721574 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="89ff8a26-ef18-4434-9601-11a47af62d20" containerName="oc" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.723880 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.725988 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.727125 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.727611 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-b8dws" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.735074 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t"] Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.736077 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.737933 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.745727 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t"] Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.826077 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-r7c5t"] Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.827258 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r7c5t" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.830443 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.830660 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-wbh7r" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.830777 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.835414 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-f9gjz"] Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.835953 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.836721 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.838290 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.848384 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-f9gjz"] Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.919477 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkvkq\" (UniqueName: \"kubernetes.io/projected/b2b10075-38e1-46a0-a69e-f076348c77c5-kube-api-access-rkvkq\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.919527 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94012e02-d626-4c3c-b57c-c62e05f7e5dd-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-5tb4t\" (UID: \"94012e02-d626-4c3c-b57c-c62e05f7e5dd\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.919812 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-metrics\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.919852 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2b10075-38e1-46a0-a69e-f076348c77c5-metrics-certs\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.919897 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-reloader\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.919918 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd54s\" (UniqueName: \"kubernetes.io/projected/94012e02-d626-4c3c-b57c-c62e05f7e5dd-kube-api-access-qd54s\") pod \"frr-k8s-webhook-server-bcc4b6f68-5tb4t\" (UID: \"94012e02-d626-4c3c-b57c-c62e05f7e5dd\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.919955 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-frr-conf\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.919992 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-frr-sockets\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:26 crc kubenswrapper[4936]: I0320 16:14:26.920024 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b2b10075-38e1-46a0-a69e-f076348c77c5-frr-startup\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021441 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-metrics\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021481 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2b10075-38e1-46a0-a69e-f076348c77c5-metrics-certs\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021506 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-memberlist\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021526 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-reloader\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021564 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd54s\" (UniqueName: \"kubernetes.io/projected/94012e02-d626-4c3c-b57c-c62e05f7e5dd-kube-api-access-qd54s\") pod \"frr-k8s-webhook-server-bcc4b6f68-5tb4t\" (UID: \"94012e02-d626-4c3c-b57c-c62e05f7e5dd\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021651 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h5fq\" (UniqueName: \"kubernetes.io/projected/bac69582-9f6c-4828-a05b-1f9bfbcefbef-kube-api-access-4h5fq\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021692 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-frr-conf\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021717 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-frr-sockets\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021745 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b2b10075-38e1-46a0-a69e-f076348c77c5-frr-startup\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021782 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0df86e9a-2b40-4705-9c88-2f24b1037d9d-cert\") pod \"controller-7bb4cc7c98-f9gjz\" (UID: \"0df86e9a-2b40-4705-9c88-2f24b1037d9d\") " pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021830 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkvkq\" (UniqueName: \"kubernetes.io/projected/b2b10075-38e1-46a0-a69e-f076348c77c5-kube-api-access-rkvkq\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021857 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/bac69582-9f6c-4828-a05b-1f9bfbcefbef-metallb-excludel2\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021880 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72km5\" (UniqueName: \"kubernetes.io/projected/0df86e9a-2b40-4705-9c88-2f24b1037d9d-kube-api-access-72km5\") pod \"controller-7bb4cc7c98-f9gjz\" (UID: \"0df86e9a-2b40-4705-9c88-2f24b1037d9d\") " pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021905 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-metrics-certs\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021930 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94012e02-d626-4c3c-b57c-c62e05f7e5dd-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-5tb4t\" (UID: \"94012e02-d626-4c3c-b57c-c62e05f7e5dd\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.021951 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0df86e9a-2b40-4705-9c88-2f24b1037d9d-metrics-certs\") pod \"controller-7bb4cc7c98-f9gjz\" (UID: \"0df86e9a-2b40-4705-9c88-2f24b1037d9d\") " pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.022055 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-reloader\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.022150 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-metrics\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.022185 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-frr-sockets\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.022465 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b2b10075-38e1-46a0-a69e-f076348c77c5-frr-conf\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.022972 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b2b10075-38e1-46a0-a69e-f076348c77c5-frr-startup\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.029212 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94012e02-d626-4c3c-b57c-c62e05f7e5dd-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-5tb4t\" (UID: \"94012e02-d626-4c3c-b57c-c62e05f7e5dd\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.029571 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2b10075-38e1-46a0-a69e-f076348c77c5-metrics-certs\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.041292 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkvkq\" (UniqueName: \"kubernetes.io/projected/b2b10075-38e1-46a0-a69e-f076348c77c5-kube-api-access-rkvkq\") pod \"frr-k8s-4fg6x\" (UID: \"b2b10075-38e1-46a0-a69e-f076348c77c5\") " pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.041916 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd54s\" (UniqueName: \"kubernetes.io/projected/94012e02-d626-4c3c-b57c-c62e05f7e5dd-kube-api-access-qd54s\") pod \"frr-k8s-webhook-server-bcc4b6f68-5tb4t\" (UID: \"94012e02-d626-4c3c-b57c-c62e05f7e5dd\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.046467 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.060913 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.132588 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-memberlist\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.132651 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h5fq\" (UniqueName: \"kubernetes.io/projected/bac69582-9f6c-4828-a05b-1f9bfbcefbef-kube-api-access-4h5fq\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.132703 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0df86e9a-2b40-4705-9c88-2f24b1037d9d-cert\") pod \"controller-7bb4cc7c98-f9gjz\" (UID: \"0df86e9a-2b40-4705-9c88-2f24b1037d9d\") " pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.132749 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/bac69582-9f6c-4828-a05b-1f9bfbcefbef-metallb-excludel2\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.132767 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72km5\" (UniqueName: \"kubernetes.io/projected/0df86e9a-2b40-4705-9c88-2f24b1037d9d-kube-api-access-72km5\") pod \"controller-7bb4cc7c98-f9gjz\" (UID: \"0df86e9a-2b40-4705-9c88-2f24b1037d9d\") " pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.132785 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-metrics-certs\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: E0320 16:14:27.132787 4936 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 20 16:14:27 crc kubenswrapper[4936]: E0320 16:14:27.132867 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-memberlist podName:bac69582-9f6c-4828-a05b-1f9bfbcefbef nodeName:}" failed. No retries permitted until 2026-03-20 16:14:27.632849233 +0000 UTC m=+818.579217048 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-memberlist") pod "speaker-r7c5t" (UID: "bac69582-9f6c-4828-a05b-1f9bfbcefbef") : secret "metallb-memberlist" not found Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.132803 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0df86e9a-2b40-4705-9c88-2f24b1037d9d-metrics-certs\") pod \"controller-7bb4cc7c98-f9gjz\" (UID: \"0df86e9a-2b40-4705-9c88-2f24b1037d9d\") " pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.134127 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/bac69582-9f6c-4828-a05b-1f9bfbcefbef-metallb-excludel2\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.136446 4936 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.139333 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-metrics-certs\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.140337 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0df86e9a-2b40-4705-9c88-2f24b1037d9d-metrics-certs\") pod \"controller-7bb4cc7c98-f9gjz\" (UID: \"0df86e9a-2b40-4705-9c88-2f24b1037d9d\") " pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.147164 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0df86e9a-2b40-4705-9c88-2f24b1037d9d-cert\") pod \"controller-7bb4cc7c98-f9gjz\" (UID: \"0df86e9a-2b40-4705-9c88-2f24b1037d9d\") " pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.152862 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h5fq\" (UniqueName: \"kubernetes.io/projected/bac69582-9f6c-4828-a05b-1f9bfbcefbef-kube-api-access-4h5fq\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.157799 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72km5\" (UniqueName: \"kubernetes.io/projected/0df86e9a-2b40-4705-9c88-2f24b1037d9d-kube-api-access-72km5\") pod \"controller-7bb4cc7c98-f9gjz\" (UID: \"0df86e9a-2b40-4705-9c88-2f24b1037d9d\") " pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.452043 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.555240 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t"] Mar 20 16:14:27 crc kubenswrapper[4936]: W0320 16:14:27.558624 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94012e02_d626_4c3c_b57c_c62e05f7e5dd.slice/crio-e1aab5443ee382fe887d2444a999e3af8cdb500528c00df0774ce057151e0af4 WatchSource:0}: Error finding container e1aab5443ee382fe887d2444a999e3af8cdb500528c00df0774ce057151e0af4: Status 404 returned error can't find the container with id e1aab5443ee382fe887d2444a999e3af8cdb500528c00df0774ce057151e0af4 Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.640884 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-memberlist\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:27 crc kubenswrapper[4936]: E0320 16:14:27.641122 4936 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 20 16:14:27 crc kubenswrapper[4936]: E0320 16:14:27.641385 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-memberlist podName:bac69582-9f6c-4828-a05b-1f9bfbcefbef nodeName:}" failed. No retries permitted until 2026-03-20 16:14:28.64136432 +0000 UTC m=+819.587732155 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-memberlist") pod "speaker-r7c5t" (UID: "bac69582-9f6c-4828-a05b-1f9bfbcefbef") : secret "metallb-memberlist" not found Mar 20 16:14:27 crc kubenswrapper[4936]: I0320 16:14:27.729790 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-f9gjz"] Mar 20 16:14:27 crc kubenswrapper[4936]: W0320 16:14:27.740169 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0df86e9a_2b40_4705_9c88_2f24b1037d9d.slice/crio-b5a0f868f5cd2f784ab4ed9e32e5aaa96a5c51748c3efe21be84ed732fe9027e WatchSource:0}: Error finding container b5a0f868f5cd2f784ab4ed9e32e5aaa96a5c51748c3efe21be84ed732fe9027e: Status 404 returned error can't find the container with id b5a0f868f5cd2f784ab4ed9e32e5aaa96a5c51748c3efe21be84ed732fe9027e Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.428037 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-f9gjz" event={"ID":"0df86e9a-2b40-4705-9c88-2f24b1037d9d","Type":"ContainerStarted","Data":"63641d6c95d50bd3ec8d5b0a65b045ef64f79733cbbc2f7efa2e599de9fa092a"} Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.428106 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-f9gjz" event={"ID":"0df86e9a-2b40-4705-9c88-2f24b1037d9d","Type":"ContainerStarted","Data":"9b856daaddec7e409b00f3428c6a8ff56f4dc1e93854265dbcb0deb4a069c624"} Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.428130 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-f9gjz" event={"ID":"0df86e9a-2b40-4705-9c88-2f24b1037d9d","Type":"ContainerStarted","Data":"b5a0f868f5cd2f784ab4ed9e32e5aaa96a5c51748c3efe21be84ed732fe9027e"} Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.428195 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.429722 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" event={"ID":"94012e02-d626-4c3c-b57c-c62e05f7e5dd","Type":"ContainerStarted","Data":"e1aab5443ee382fe887d2444a999e3af8cdb500528c00df0774ce057151e0af4"} Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.431131 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerStarted","Data":"5e8ac349a3fc5864982e53ff0d979d42d1f6165b95ecf85215fab9014eb4eda6"} Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.460215 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-f9gjz" podStartSLOduration=2.46018677 podStartE2EDuration="2.46018677s" podCreationTimestamp="2026-03-20 16:14:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:14:28.456667958 +0000 UTC m=+819.403035833" watchObservedRunningTime="2026-03-20 16:14:28.46018677 +0000 UTC m=+819.406554625" Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.654134 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-memberlist\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.663373 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bac69582-9f6c-4828-a05b-1f9bfbcefbef-memberlist\") pod \"speaker-r7c5t\" (UID: \"bac69582-9f6c-4828-a05b-1f9bfbcefbef\") " pod="metallb-system/speaker-r7c5t" Mar 20 16:14:28 crc kubenswrapper[4936]: I0320 16:14:28.945076 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r7c5t" Mar 20 16:14:29 crc kubenswrapper[4936]: I0320 16:14:29.443657 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r7c5t" event={"ID":"bac69582-9f6c-4828-a05b-1f9bfbcefbef","Type":"ContainerStarted","Data":"e29d4a51b187579607ebbaaedc650842ea297aee23dc968c035f7aefdfea828d"} Mar 20 16:14:29 crc kubenswrapper[4936]: I0320 16:14:29.443689 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r7c5t" event={"ID":"bac69582-9f6c-4828-a05b-1f9bfbcefbef","Type":"ContainerStarted","Data":"afceb34a720a59930c81051909797098e056f3b28cbc990adfa373b0103bc2e3"} Mar 20 16:14:30 crc kubenswrapper[4936]: I0320 16:14:30.454859 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r7c5t" event={"ID":"bac69582-9f6c-4828-a05b-1f9bfbcefbef","Type":"ContainerStarted","Data":"eb8c222898ff328e1226bea5f5938dbee5ef3026b3824776d21a8e2b4d717f4b"} Mar 20 16:14:30 crc kubenswrapper[4936]: I0320 16:14:30.456287 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-r7c5t" Mar 20 16:14:30 crc kubenswrapper[4936]: I0320 16:14:30.488978 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-r7c5t" podStartSLOduration=4.488964001 podStartE2EDuration="4.488964001s" podCreationTimestamp="2026-03-20 16:14:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:14:30.487737726 +0000 UTC m=+821.434105541" watchObservedRunningTime="2026-03-20 16:14:30.488964001 +0000 UTC m=+821.435331816" Mar 20 16:14:35 crc kubenswrapper[4936]: I0320 16:14:35.486757 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" event={"ID":"94012e02-d626-4c3c-b57c-c62e05f7e5dd","Type":"ContainerStarted","Data":"f8990e4828c787c02f49b65543784e71f7c13cd19043a741f4a2ca133354eeaa"} Mar 20 16:14:35 crc kubenswrapper[4936]: I0320 16:14:35.487274 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:35 crc kubenswrapper[4936]: I0320 16:14:35.488488 4936 generic.go:334] "Generic (PLEG): container finished" podID="b2b10075-38e1-46a0-a69e-f076348c77c5" containerID="3936f989da0ccea8571a7d59650b02a0a6d36cc9e7d58b3138936fa4a6c70eca" exitCode=0 Mar 20 16:14:35 crc kubenswrapper[4936]: I0320 16:14:35.488593 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerDied","Data":"3936f989da0ccea8571a7d59650b02a0a6d36cc9e7d58b3138936fa4a6c70eca"} Mar 20 16:14:35 crc kubenswrapper[4936]: I0320 16:14:35.511528 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" podStartSLOduration=1.851577553 podStartE2EDuration="9.51150637s" podCreationTimestamp="2026-03-20 16:14:26 +0000 UTC" firstStartedPulling="2026-03-20 16:14:27.563780646 +0000 UTC m=+818.510148501" lastFinishedPulling="2026-03-20 16:14:35.223709503 +0000 UTC m=+826.170077318" observedRunningTime="2026-03-20 16:14:35.505201006 +0000 UTC m=+826.451568831" watchObservedRunningTime="2026-03-20 16:14:35.51150637 +0000 UTC m=+826.457874185" Mar 20 16:14:36 crc kubenswrapper[4936]: I0320 16:14:36.497351 4936 generic.go:334] "Generic (PLEG): container finished" podID="b2b10075-38e1-46a0-a69e-f076348c77c5" containerID="2bd909cead7ea0113eec330834d3e98a0067d391c711410b3ee930682fde330e" exitCode=0 Mar 20 16:14:36 crc kubenswrapper[4936]: I0320 16:14:36.497463 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerDied","Data":"2bd909cead7ea0113eec330834d3e98a0067d391c711410b3ee930682fde330e"} Mar 20 16:14:37 crc kubenswrapper[4936]: I0320 16:14:37.455956 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-f9gjz" Mar 20 16:14:37 crc kubenswrapper[4936]: I0320 16:14:37.504148 4936 generic.go:334] "Generic (PLEG): container finished" podID="b2b10075-38e1-46a0-a69e-f076348c77c5" containerID="f86348eb4fef59c3a1b0ee88fb5209c937bda4b1b557573aef8d3065390a88da" exitCode=0 Mar 20 16:14:37 crc kubenswrapper[4936]: I0320 16:14:37.504201 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerDied","Data":"f86348eb4fef59c3a1b0ee88fb5209c937bda4b1b557573aef8d3065390a88da"} Mar 20 16:14:38 crc kubenswrapper[4936]: I0320 16:14:38.549428 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerStarted","Data":"87b01bfa2db9c76a3b9b22f89f48f9ded75b942c9d3450d551851208c49db892"} Mar 20 16:14:38 crc kubenswrapper[4936]: I0320 16:14:38.549473 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerStarted","Data":"b68e337d1de5af51be1684571f70f386acb8f2560c14356d765426f1cb57a907"} Mar 20 16:14:38 crc kubenswrapper[4936]: I0320 16:14:38.549483 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerStarted","Data":"21a597cd77ffa95bb53d11098aefc402d8098b16fa05e5fccabc6977237b18bc"} Mar 20 16:14:38 crc kubenswrapper[4936]: I0320 16:14:38.549491 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerStarted","Data":"fec38b73bdd15059355aca8b781446b34089c777277363b7a3e9c7bbda50d3f8"} Mar 20 16:14:38 crc kubenswrapper[4936]: I0320 16:14:38.549500 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerStarted","Data":"1926074b73c2c61857dc49857d1b5371fce1b422a06f383052919755ed61854a"} Mar 20 16:14:39 crc kubenswrapper[4936]: I0320 16:14:39.562507 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4fg6x" event={"ID":"b2b10075-38e1-46a0-a69e-f076348c77c5","Type":"ContainerStarted","Data":"d508f3489410508baea83511d545a59a34600e1bcccf46945898275fa81c3a52"} Mar 20 16:14:39 crc kubenswrapper[4936]: I0320 16:14:39.562723 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:39 crc kubenswrapper[4936]: I0320 16:14:39.601362 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-4fg6x" podStartSLOduration=5.882956224 podStartE2EDuration="13.601340485s" podCreationTimestamp="2026-03-20 16:14:26 +0000 UTC" firstStartedPulling="2026-03-20 16:14:27.521664708 +0000 UTC m=+818.468032563" lastFinishedPulling="2026-03-20 16:14:35.240049009 +0000 UTC m=+826.186416824" observedRunningTime="2026-03-20 16:14:39.592323082 +0000 UTC m=+830.538690937" watchObservedRunningTime="2026-03-20 16:14:39.601340485 +0000 UTC m=+830.547708330" Mar 20 16:14:42 crc kubenswrapper[4936]: I0320 16:14:42.046800 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:42 crc kubenswrapper[4936]: I0320 16:14:42.094306 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:47 crc kubenswrapper[4936]: I0320 16:14:47.050589 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-4fg6x" Mar 20 16:14:47 crc kubenswrapper[4936]: I0320 16:14:47.075334 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-5tb4t" Mar 20 16:14:48 crc kubenswrapper[4936]: I0320 16:14:48.950567 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-r7c5t" Mar 20 16:14:50 crc kubenswrapper[4936]: I0320 16:14:50.455743 4936 scope.go:117] "RemoveContainer" containerID="e5f3de3dfb82abd886df377c901d821179552144b341c3b1fa6b7f45bc7753e6" Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.112563 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-76njm"] Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.113580 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-76njm" Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.117339 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.117594 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-sks7j" Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.119898 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.138444 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-76njm"] Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.191032 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gskj8\" (UniqueName: \"kubernetes.io/projected/566f7141-5068-4e6e-9d06-bf61426d5365-kube-api-access-gskj8\") pod \"openstack-operator-index-76njm\" (UID: \"566f7141-5068-4e6e-9d06-bf61426d5365\") " pod="openstack-operators/openstack-operator-index-76njm" Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.292635 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gskj8\" (UniqueName: \"kubernetes.io/projected/566f7141-5068-4e6e-9d06-bf61426d5365-kube-api-access-gskj8\") pod \"openstack-operator-index-76njm\" (UID: \"566f7141-5068-4e6e-9d06-bf61426d5365\") " pod="openstack-operators/openstack-operator-index-76njm" Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.331326 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gskj8\" (UniqueName: \"kubernetes.io/projected/566f7141-5068-4e6e-9d06-bf61426d5365-kube-api-access-gskj8\") pod \"openstack-operator-index-76njm\" (UID: \"566f7141-5068-4e6e-9d06-bf61426d5365\") " pod="openstack-operators/openstack-operator-index-76njm" Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.433401 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-76njm" Mar 20 16:14:53 crc kubenswrapper[4936]: I0320 16:14:53.850879 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-76njm"] Mar 20 16:14:53 crc kubenswrapper[4936]: W0320 16:14:53.858498 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod566f7141_5068_4e6e_9d06_bf61426d5365.slice/crio-09bcb920cc07d4230fae6bc407aa17c0b0b135d69567fdde6c7321bc533de522 WatchSource:0}: Error finding container 09bcb920cc07d4230fae6bc407aa17c0b0b135d69567fdde6c7321bc533de522: Status 404 returned error can't find the container with id 09bcb920cc07d4230fae6bc407aa17c0b0b135d69567fdde6c7321bc533de522 Mar 20 16:14:54 crc kubenswrapper[4936]: I0320 16:14:54.677643 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-76njm" event={"ID":"566f7141-5068-4e6e-9d06-bf61426d5365","Type":"ContainerStarted","Data":"09bcb920cc07d4230fae6bc407aa17c0b0b135d69567fdde6c7321bc533de522"} Mar 20 16:14:56 crc kubenswrapper[4936]: I0320 16:14:56.496271 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-76njm"] Mar 20 16:14:56 crc kubenswrapper[4936]: I0320 16:14:56.694120 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-76njm" event={"ID":"566f7141-5068-4e6e-9d06-bf61426d5365","Type":"ContainerStarted","Data":"83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786"} Mar 20 16:14:56 crc kubenswrapper[4936]: I0320 16:14:56.715114 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-76njm" podStartSLOduration=1.325817736 podStartE2EDuration="3.715086642s" podCreationTimestamp="2026-03-20 16:14:53 +0000 UTC" firstStartedPulling="2026-03-20 16:14:53.863462685 +0000 UTC m=+844.809830500" lastFinishedPulling="2026-03-20 16:14:56.252731591 +0000 UTC m=+847.199099406" observedRunningTime="2026-03-20 16:14:56.708722736 +0000 UTC m=+847.655090591" watchObservedRunningTime="2026-03-20 16:14:56.715086642 +0000 UTC m=+847.661454487" Mar 20 16:14:57 crc kubenswrapper[4936]: I0320 16:14:57.107274 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-c4nw2"] Mar 20 16:14:57 crc kubenswrapper[4936]: I0320 16:14:57.108959 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-c4nw2" Mar 20 16:14:57 crc kubenswrapper[4936]: I0320 16:14:57.119934 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-c4nw2"] Mar 20 16:14:57 crc kubenswrapper[4936]: I0320 16:14:57.244416 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slwkg\" (UniqueName: \"kubernetes.io/projected/0401145a-6ad6-44e3-b2ab-ab333cc2c118-kube-api-access-slwkg\") pod \"openstack-operator-index-c4nw2\" (UID: \"0401145a-6ad6-44e3-b2ab-ab333cc2c118\") " pod="openstack-operators/openstack-operator-index-c4nw2" Mar 20 16:14:57 crc kubenswrapper[4936]: I0320 16:14:57.345420 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slwkg\" (UniqueName: \"kubernetes.io/projected/0401145a-6ad6-44e3-b2ab-ab333cc2c118-kube-api-access-slwkg\") pod \"openstack-operator-index-c4nw2\" (UID: \"0401145a-6ad6-44e3-b2ab-ab333cc2c118\") " pod="openstack-operators/openstack-operator-index-c4nw2" Mar 20 16:14:57 crc kubenswrapper[4936]: I0320 16:14:57.378756 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slwkg\" (UniqueName: \"kubernetes.io/projected/0401145a-6ad6-44e3-b2ab-ab333cc2c118-kube-api-access-slwkg\") pod \"openstack-operator-index-c4nw2\" (UID: \"0401145a-6ad6-44e3-b2ab-ab333cc2c118\") " pod="openstack-operators/openstack-operator-index-c4nw2" Mar 20 16:14:57 crc kubenswrapper[4936]: I0320 16:14:57.438470 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-c4nw2" Mar 20 16:14:57 crc kubenswrapper[4936]: I0320 16:14:57.701031 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-76njm" podUID="566f7141-5068-4e6e-9d06-bf61426d5365" containerName="registry-server" containerID="cri-o://83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786" gracePeriod=2 Mar 20 16:14:57 crc kubenswrapper[4936]: I0320 16:14:57.898821 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-c4nw2"] Mar 20 16:14:57 crc kubenswrapper[4936]: W0320 16:14:57.912776 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0401145a_6ad6_44e3_b2ab_ab333cc2c118.slice/crio-9a64428168a304c7f9f144a775ad17e7a3b9af8e77ff5b516bae333d3db0594c WatchSource:0}: Error finding container 9a64428168a304c7f9f144a775ad17e7a3b9af8e77ff5b516bae333d3db0594c: Status 404 returned error can't find the container with id 9a64428168a304c7f9f144a775ad17e7a3b9af8e77ff5b516bae333d3db0594c Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.003682 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-76njm" Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.062850 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gskj8\" (UniqueName: \"kubernetes.io/projected/566f7141-5068-4e6e-9d06-bf61426d5365-kube-api-access-gskj8\") pod \"566f7141-5068-4e6e-9d06-bf61426d5365\" (UID: \"566f7141-5068-4e6e-9d06-bf61426d5365\") " Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.069199 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/566f7141-5068-4e6e-9d06-bf61426d5365-kube-api-access-gskj8" (OuterVolumeSpecName: "kube-api-access-gskj8") pod "566f7141-5068-4e6e-9d06-bf61426d5365" (UID: "566f7141-5068-4e6e-9d06-bf61426d5365"). InnerVolumeSpecName "kube-api-access-gskj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.164437 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gskj8\" (UniqueName: \"kubernetes.io/projected/566f7141-5068-4e6e-9d06-bf61426d5365-kube-api-access-gskj8\") on node \"crc\" DevicePath \"\"" Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.711155 4936 generic.go:334] "Generic (PLEG): container finished" podID="566f7141-5068-4e6e-9d06-bf61426d5365" containerID="83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786" exitCode=0 Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.711235 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-76njm" Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.711227 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-76njm" event={"ID":"566f7141-5068-4e6e-9d06-bf61426d5365","Type":"ContainerDied","Data":"83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786"} Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.713075 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-76njm" event={"ID":"566f7141-5068-4e6e-9d06-bf61426d5365","Type":"ContainerDied","Data":"09bcb920cc07d4230fae6bc407aa17c0b0b135d69567fdde6c7321bc533de522"} Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.713124 4936 scope.go:117] "RemoveContainer" containerID="83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786" Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.714923 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-c4nw2" event={"ID":"0401145a-6ad6-44e3-b2ab-ab333cc2c118","Type":"ContainerStarted","Data":"2b3ba9d3777c22c3b0c1a79d2946006f4aa61af60d0c19ce7812a6b30ed4a368"} Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.714981 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-c4nw2" event={"ID":"0401145a-6ad6-44e3-b2ab-ab333cc2c118","Type":"ContainerStarted","Data":"9a64428168a304c7f9f144a775ad17e7a3b9af8e77ff5b516bae333d3db0594c"} Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.741311 4936 scope.go:117] "RemoveContainer" containerID="83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786" Mar 20 16:14:58 crc kubenswrapper[4936]: E0320 16:14:58.742324 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786\": container with ID starting with 83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786 not found: ID does not exist" containerID="83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786" Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.742365 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786"} err="failed to get container status \"83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786\": rpc error: code = NotFound desc = could not find container \"83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786\": container with ID starting with 83263de83e082f7e2b2f11761e1a436c4c7ec5cfcb4e13d719a78f02ccd65786 not found: ID does not exist" Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.755957 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-c4nw2" podStartSLOduration=1.702498454 podStartE2EDuration="1.755931481s" podCreationTimestamp="2026-03-20 16:14:57 +0000 UTC" firstStartedPulling="2026-03-20 16:14:57.932179991 +0000 UTC m=+848.878547816" lastFinishedPulling="2026-03-20 16:14:57.985613028 +0000 UTC m=+848.931980843" observedRunningTime="2026-03-20 16:14:58.740320133 +0000 UTC m=+849.686688008" watchObservedRunningTime="2026-03-20 16:14:58.755931481 +0000 UTC m=+849.702299316" Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.761143 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-76njm"] Mar 20 16:14:58 crc kubenswrapper[4936]: I0320 16:14:58.766079 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-76njm"] Mar 20 16:14:59 crc kubenswrapper[4936]: I0320 16:14:59.876075 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="566f7141-5068-4e6e-9d06-bf61426d5365" path="/var/lib/kubelet/pods/566f7141-5068-4e6e-9d06-bf61426d5365/volumes" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.133107 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx"] Mar 20 16:15:00 crc kubenswrapper[4936]: E0320 16:15:00.133470 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566f7141-5068-4e6e-9d06-bf61426d5365" containerName="registry-server" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.133499 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="566f7141-5068-4e6e-9d06-bf61426d5365" containerName="registry-server" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.133775 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="566f7141-5068-4e6e-9d06-bf61426d5365" containerName="registry-server" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.134435 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.140876 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx"] Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.150449 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.151957 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.190087 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-config-volume\") pod \"collect-profiles-29567055-q6qkx\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.190160 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx467\" (UniqueName: \"kubernetes.io/projected/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-kube-api-access-bx467\") pod \"collect-profiles-29567055-q6qkx\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.190194 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-secret-volume\") pod \"collect-profiles-29567055-q6qkx\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.291596 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx467\" (UniqueName: \"kubernetes.io/projected/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-kube-api-access-bx467\") pod \"collect-profiles-29567055-q6qkx\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.292291 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-secret-volume\") pod \"collect-profiles-29567055-q6qkx\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.293393 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-config-volume\") pod \"collect-profiles-29567055-q6qkx\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.294098 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-config-volume\") pod \"collect-profiles-29567055-q6qkx\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.315253 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-secret-volume\") pod \"collect-profiles-29567055-q6qkx\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.321805 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx467\" (UniqueName: \"kubernetes.io/projected/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-kube-api-access-bx467\") pod \"collect-profiles-29567055-q6qkx\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.451802 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:00 crc kubenswrapper[4936]: I0320 16:15:00.985656 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx"] Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.502752 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-25c56"] Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.504358 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.514606 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-utilities\") pod \"redhat-marketplace-25c56\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.514663 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-catalog-content\") pod \"redhat-marketplace-25c56\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.514684 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x29v6\" (UniqueName: \"kubernetes.io/projected/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-kube-api-access-x29v6\") pod \"redhat-marketplace-25c56\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.515598 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25c56"] Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.615584 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-utilities\") pod \"redhat-marketplace-25c56\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.615634 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-catalog-content\") pod \"redhat-marketplace-25c56\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.615656 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x29v6\" (UniqueName: \"kubernetes.io/projected/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-kube-api-access-x29v6\") pod \"redhat-marketplace-25c56\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.616088 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-utilities\") pod \"redhat-marketplace-25c56\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.616191 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-catalog-content\") pod \"redhat-marketplace-25c56\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.634839 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x29v6\" (UniqueName: \"kubernetes.io/projected/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-kube-api-access-x29v6\") pod \"redhat-marketplace-25c56\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.736383 4936 generic.go:334] "Generic (PLEG): container finished" podID="54cecdd8-6e0a-4e9f-92fb-7ed277bc2872" containerID="7cf4b578a3d90238a9b1ae1c90fb41e3eed72cc93967fd55b192cea35b117a1a" exitCode=0 Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.736434 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" event={"ID":"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872","Type":"ContainerDied","Data":"7cf4b578a3d90238a9b1ae1c90fb41e3eed72cc93967fd55b192cea35b117a1a"} Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.736466 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" event={"ID":"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872","Type":"ContainerStarted","Data":"babf85c03a300b2f47d2e09c74b736341300dc98ecab3d1b21cc1218eceadebf"} Mar 20 16:15:01 crc kubenswrapper[4936]: I0320 16:15:01.827693 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:02 crc kubenswrapper[4936]: I0320 16:15:02.239580 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25c56"] Mar 20 16:15:02 crc kubenswrapper[4936]: I0320 16:15:02.752438 4936 generic.go:334] "Generic (PLEG): container finished" podID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerID="cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760" exitCode=0 Mar 20 16:15:02 crc kubenswrapper[4936]: I0320 16:15:02.752604 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25c56" event={"ID":"23d93095-0ea4-4c2c-ac82-cbf756a2e82e","Type":"ContainerDied","Data":"cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760"} Mar 20 16:15:02 crc kubenswrapper[4936]: I0320 16:15:02.752653 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25c56" event={"ID":"23d93095-0ea4-4c2c-ac82-cbf756a2e82e","Type":"ContainerStarted","Data":"e74c6ca8a212327def697e50c9f3f0e3e70b0689da9c933608d08391a081d5de"} Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.011242 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.045327 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx467\" (UniqueName: \"kubernetes.io/projected/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-kube-api-access-bx467\") pod \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.045438 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-secret-volume\") pod \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.045494 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-config-volume\") pod \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\" (UID: \"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872\") " Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.046656 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-config-volume" (OuterVolumeSpecName: "config-volume") pod "54cecdd8-6e0a-4e9f-92fb-7ed277bc2872" (UID: "54cecdd8-6e0a-4e9f-92fb-7ed277bc2872"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.053882 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-kube-api-access-bx467" (OuterVolumeSpecName: "kube-api-access-bx467") pod "54cecdd8-6e0a-4e9f-92fb-7ed277bc2872" (UID: "54cecdd8-6e0a-4e9f-92fb-7ed277bc2872"). InnerVolumeSpecName "kube-api-access-bx467". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.056676 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "54cecdd8-6e0a-4e9f-92fb-7ed277bc2872" (UID: "54cecdd8-6e0a-4e9f-92fb-7ed277bc2872"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.147959 4936 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.148062 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx467\" (UniqueName: \"kubernetes.io/projected/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-kube-api-access-bx467\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.148089 4936 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54cecdd8-6e0a-4e9f-92fb-7ed277bc2872-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.760473 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25c56" event={"ID":"23d93095-0ea4-4c2c-ac82-cbf756a2e82e","Type":"ContainerStarted","Data":"600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db"} Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.762638 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" event={"ID":"54cecdd8-6e0a-4e9f-92fb-7ed277bc2872","Type":"ContainerDied","Data":"babf85c03a300b2f47d2e09c74b736341300dc98ecab3d1b21cc1218eceadebf"} Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.762679 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="babf85c03a300b2f47d2e09c74b736341300dc98ecab3d1b21cc1218eceadebf" Mar 20 16:15:03 crc kubenswrapper[4936]: I0320 16:15:03.762661 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567055-q6qkx" Mar 20 16:15:04 crc kubenswrapper[4936]: I0320 16:15:04.771489 4936 generic.go:334] "Generic (PLEG): container finished" podID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerID="600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db" exitCode=0 Mar 20 16:15:04 crc kubenswrapper[4936]: I0320 16:15:04.771565 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25c56" event={"ID":"23d93095-0ea4-4c2c-ac82-cbf756a2e82e","Type":"ContainerDied","Data":"600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db"} Mar 20 16:15:05 crc kubenswrapper[4936]: I0320 16:15:05.781530 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25c56" event={"ID":"23d93095-0ea4-4c2c-ac82-cbf756a2e82e","Type":"ContainerStarted","Data":"a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40"} Mar 20 16:15:05 crc kubenswrapper[4936]: I0320 16:15:05.799395 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-25c56" podStartSLOduration=1.993382187 podStartE2EDuration="4.799367538s" podCreationTimestamp="2026-03-20 16:15:01 +0000 UTC" firstStartedPulling="2026-03-20 16:15:02.754280455 +0000 UTC m=+853.700648270" lastFinishedPulling="2026-03-20 16:15:05.560265806 +0000 UTC m=+856.506633621" observedRunningTime="2026-03-20 16:15:05.797197315 +0000 UTC m=+856.743565150" watchObservedRunningTime="2026-03-20 16:15:05.799367538 +0000 UTC m=+856.745735393" Mar 20 16:15:07 crc kubenswrapper[4936]: I0320 16:15:07.438834 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-c4nw2" Mar 20 16:15:07 crc kubenswrapper[4936]: I0320 16:15:07.439200 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-c4nw2" Mar 20 16:15:07 crc kubenswrapper[4936]: I0320 16:15:07.473197 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-c4nw2" Mar 20 16:15:07 crc kubenswrapper[4936]: I0320 16:15:07.815422 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-c4nw2" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.752896 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz"] Mar 20 16:15:10 crc kubenswrapper[4936]: E0320 16:15:10.753644 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cecdd8-6e0a-4e9f-92fb-7ed277bc2872" containerName="collect-profiles" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.753664 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cecdd8-6e0a-4e9f-92fb-7ed277bc2872" containerName="collect-profiles" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.753900 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cecdd8-6e0a-4e9f-92fb-7ed277bc2872" containerName="collect-profiles" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.755334 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.758513 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-m9qhd" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.760136 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz"] Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.880798 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkcmx\" (UniqueName: \"kubernetes.io/projected/cada20db-602e-45c5-be95-65c083299de3-kube-api-access-gkcmx\") pod \"5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.880860 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-util\") pod \"5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.880898 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-bundle\") pod \"5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.982866 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkcmx\" (UniqueName: \"kubernetes.io/projected/cada20db-602e-45c5-be95-65c083299de3-kube-api-access-gkcmx\") pod \"5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.982981 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-util\") pod \"5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.983062 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-bundle\") pod \"5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.984020 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-util\") pod \"5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:10 crc kubenswrapper[4936]: I0320 16:15:10.984032 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-bundle\") pod \"5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:11 crc kubenswrapper[4936]: I0320 16:15:11.014932 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkcmx\" (UniqueName: \"kubernetes.io/projected/cada20db-602e-45c5-be95-65c083299de3-kube-api-access-gkcmx\") pod \"5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:11 crc kubenswrapper[4936]: I0320 16:15:11.079458 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:11 crc kubenswrapper[4936]: I0320 16:15:11.527859 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz"] Mar 20 16:15:11 crc kubenswrapper[4936]: I0320 16:15:11.823946 4936 generic.go:334] "Generic (PLEG): container finished" podID="cada20db-602e-45c5-be95-65c083299de3" containerID="7941eb233f8f21031f7b7781033cd34ba58e324500580069dc807d9b9e3b0266" exitCode=0 Mar 20 16:15:11 crc kubenswrapper[4936]: I0320 16:15:11.824031 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" event={"ID":"cada20db-602e-45c5-be95-65c083299de3","Type":"ContainerDied","Data":"7941eb233f8f21031f7b7781033cd34ba58e324500580069dc807d9b9e3b0266"} Mar 20 16:15:11 crc kubenswrapper[4936]: I0320 16:15:11.824112 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" event={"ID":"cada20db-602e-45c5-be95-65c083299de3","Type":"ContainerStarted","Data":"a67b0c0e10c4eaf55fa5a02240f4a9bb689c45672b22d3f96c2be9247f7b5204"} Mar 20 16:15:11 crc kubenswrapper[4936]: I0320 16:15:11.827842 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:11 crc kubenswrapper[4936]: I0320 16:15:11.827894 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:11 crc kubenswrapper[4936]: I0320 16:15:11.874893 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:12 crc kubenswrapper[4936]: I0320 16:15:12.836311 4936 generic.go:334] "Generic (PLEG): container finished" podID="cada20db-602e-45c5-be95-65c083299de3" containerID="6d9474af7a3fe68b1e99d5365b53e77a9a1a3cf3c5eb18b65d96334687054a23" exitCode=0 Mar 20 16:15:12 crc kubenswrapper[4936]: I0320 16:15:12.836398 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" event={"ID":"cada20db-602e-45c5-be95-65c083299de3","Type":"ContainerDied","Data":"6d9474af7a3fe68b1e99d5365b53e77a9a1a3cf3c5eb18b65d96334687054a23"} Mar 20 16:15:12 crc kubenswrapper[4936]: I0320 16:15:12.895651 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:13 crc kubenswrapper[4936]: I0320 16:15:13.854678 4936 generic.go:334] "Generic (PLEG): container finished" podID="cada20db-602e-45c5-be95-65c083299de3" containerID="cbebc24dfd8f4c713e0c479fbe6a648ba42a793ccc74b9acb25022ab4503a61d" exitCode=0 Mar 20 16:15:13 crc kubenswrapper[4936]: I0320 16:15:13.869207 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" event={"ID":"cada20db-602e-45c5-be95-65c083299de3","Type":"ContainerDied","Data":"cbebc24dfd8f4c713e0c479fbe6a648ba42a793ccc74b9acb25022ab4503a61d"} Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.102733 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25c56"] Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.105030 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-25c56" podUID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerName="registry-server" containerID="cri-o://a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40" gracePeriod=2 Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.118392 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.240146 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkcmx\" (UniqueName: \"kubernetes.io/projected/cada20db-602e-45c5-be95-65c083299de3-kube-api-access-gkcmx\") pod \"cada20db-602e-45c5-be95-65c083299de3\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.240223 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-bundle\") pod \"cada20db-602e-45c5-be95-65c083299de3\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.240347 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-util\") pod \"cada20db-602e-45c5-be95-65c083299de3\" (UID: \"cada20db-602e-45c5-be95-65c083299de3\") " Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.243743 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-bundle" (OuterVolumeSpecName: "bundle") pod "cada20db-602e-45c5-be95-65c083299de3" (UID: "cada20db-602e-45c5-be95-65c083299de3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.248818 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cada20db-602e-45c5-be95-65c083299de3-kube-api-access-gkcmx" (OuterVolumeSpecName: "kube-api-access-gkcmx") pod "cada20db-602e-45c5-be95-65c083299de3" (UID: "cada20db-602e-45c5-be95-65c083299de3"). InnerVolumeSpecName "kube-api-access-gkcmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.257432 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-util" (OuterVolumeSpecName: "util") pod "cada20db-602e-45c5-be95-65c083299de3" (UID: "cada20db-602e-45c5-be95-65c083299de3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.341829 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkcmx\" (UniqueName: \"kubernetes.io/projected/cada20db-602e-45c5-be95-65c083299de3-kube-api-access-gkcmx\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.341874 4936 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.341893 4936 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cada20db-602e-45c5-be95-65c083299de3-util\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.458049 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.543699 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-utilities\") pod \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.543775 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x29v6\" (UniqueName: \"kubernetes.io/projected/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-kube-api-access-x29v6\") pod \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.543886 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-catalog-content\") pod \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\" (UID: \"23d93095-0ea4-4c2c-ac82-cbf756a2e82e\") " Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.544631 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-utilities" (OuterVolumeSpecName: "utilities") pod "23d93095-0ea4-4c2c-ac82-cbf756a2e82e" (UID: "23d93095-0ea4-4c2c-ac82-cbf756a2e82e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.547839 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-kube-api-access-x29v6" (OuterVolumeSpecName: "kube-api-access-x29v6") pod "23d93095-0ea4-4c2c-ac82-cbf756a2e82e" (UID: "23d93095-0ea4-4c2c-ac82-cbf756a2e82e"). InnerVolumeSpecName "kube-api-access-x29v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.570310 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23d93095-0ea4-4c2c-ac82-cbf756a2e82e" (UID: "23d93095-0ea4-4c2c-ac82-cbf756a2e82e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.645064 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.645107 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x29v6\" (UniqueName: \"kubernetes.io/projected/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-kube-api-access-x29v6\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.645123 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23d93095-0ea4-4c2c-ac82-cbf756a2e82e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.867085 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" event={"ID":"cada20db-602e-45c5-be95-65c083299de3","Type":"ContainerDied","Data":"a67b0c0e10c4eaf55fa5a02240f4a9bb689c45672b22d3f96c2be9247f7b5204"} Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.867350 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a67b0c0e10c4eaf55fa5a02240f4a9bb689c45672b22d3f96c2be9247f7b5204" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.867099 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.868941 4936 generic.go:334] "Generic (PLEG): container finished" podID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerID="a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40" exitCode=0 Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.868975 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25c56" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.868977 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25c56" event={"ID":"23d93095-0ea4-4c2c-ac82-cbf756a2e82e","Type":"ContainerDied","Data":"a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40"} Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.869100 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25c56" event={"ID":"23d93095-0ea4-4c2c-ac82-cbf756a2e82e","Type":"ContainerDied","Data":"e74c6ca8a212327def697e50c9f3f0e3e70b0689da9c933608d08391a081d5de"} Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.869137 4936 scope.go:117] "RemoveContainer" containerID="a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.886998 4936 scope.go:117] "RemoveContainer" containerID="600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.905259 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25c56"] Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.909777 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-25c56"] Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.910741 4936 scope.go:117] "RemoveContainer" containerID="cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.926607 4936 scope.go:117] "RemoveContainer" containerID="a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40" Mar 20 16:15:15 crc kubenswrapper[4936]: E0320 16:15:15.926977 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40\": container with ID starting with a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40 not found: ID does not exist" containerID="a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.927014 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40"} err="failed to get container status \"a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40\": rpc error: code = NotFound desc = could not find container \"a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40\": container with ID starting with a67dfae837b0201d132f4445f742efeed9c35c2f8b51e0c02a47bfd261cdeb40 not found: ID does not exist" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.927039 4936 scope.go:117] "RemoveContainer" containerID="600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db" Mar 20 16:15:15 crc kubenswrapper[4936]: E0320 16:15:15.927218 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db\": container with ID starting with 600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db not found: ID does not exist" containerID="600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.927244 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db"} err="failed to get container status \"600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db\": rpc error: code = NotFound desc = could not find container \"600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db\": container with ID starting with 600c5a840e2c69cf045d7699e9d759698ea734f558743d943190aeb6d8eee6db not found: ID does not exist" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.927256 4936 scope.go:117] "RemoveContainer" containerID="cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760" Mar 20 16:15:15 crc kubenswrapper[4936]: E0320 16:15:15.927429 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760\": container with ID starting with cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760 not found: ID does not exist" containerID="cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760" Mar 20 16:15:15 crc kubenswrapper[4936]: I0320 16:15:15.927451 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760"} err="failed to get container status \"cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760\": rpc error: code = NotFound desc = could not find container \"cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760\": container with ID starting with cbcd036783327075957270574e8f3735f9b25af99378022176cd24c057ddb760 not found: ID does not exist" Mar 20 16:15:17 crc kubenswrapper[4936]: I0320 16:15:17.862380 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" path="/var/lib/kubelet/pods/23d93095-0ea4-4c2c-ac82-cbf756a2e82e/volumes" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.009276 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9"] Mar 20 16:15:19 crc kubenswrapper[4936]: E0320 16:15:19.010263 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerName="extract-utilities" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.010387 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerName="extract-utilities" Mar 20 16:15:19 crc kubenswrapper[4936]: E0320 16:15:19.010470 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cada20db-602e-45c5-be95-65c083299de3" containerName="pull" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.010535 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cada20db-602e-45c5-be95-65c083299de3" containerName="pull" Mar 20 16:15:19 crc kubenswrapper[4936]: E0320 16:15:19.010632 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cada20db-602e-45c5-be95-65c083299de3" containerName="extract" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.010704 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cada20db-602e-45c5-be95-65c083299de3" containerName="extract" Mar 20 16:15:19 crc kubenswrapper[4936]: E0320 16:15:19.010793 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerName="extract-content" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.010863 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerName="extract-content" Mar 20 16:15:19 crc kubenswrapper[4936]: E0320 16:15:19.010929 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerName="registry-server" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.011000 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerName="registry-server" Mar 20 16:15:19 crc kubenswrapper[4936]: E0320 16:15:19.011078 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cada20db-602e-45c5-be95-65c083299de3" containerName="util" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.011150 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cada20db-602e-45c5-be95-65c083299de3" containerName="util" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.011362 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cada20db-602e-45c5-be95-65c083299de3" containerName="extract" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.011442 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="23d93095-0ea4-4c2c-ac82-cbf756a2e82e" containerName="registry-server" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.011970 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.013860 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-hf5fs" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.034064 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9"] Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.089081 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dqzg\" (UniqueName: \"kubernetes.io/projected/c4efa574-8274-4c0b-9463-d076e31d2a7e-kube-api-access-4dqzg\") pod \"openstack-operator-controller-init-658f6bdff6-vrsv9\" (UID: \"c4efa574-8274-4c0b-9463-d076e31d2a7e\") " pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.190645 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dqzg\" (UniqueName: \"kubernetes.io/projected/c4efa574-8274-4c0b-9463-d076e31d2a7e-kube-api-access-4dqzg\") pod \"openstack-operator-controller-init-658f6bdff6-vrsv9\" (UID: \"c4efa574-8274-4c0b-9463-d076e31d2a7e\") " pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.210992 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dqzg\" (UniqueName: \"kubernetes.io/projected/c4efa574-8274-4c0b-9463-d076e31d2a7e-kube-api-access-4dqzg\") pod \"openstack-operator-controller-init-658f6bdff6-vrsv9\" (UID: \"c4efa574-8274-4c0b-9463-d076e31d2a7e\") " pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.330571 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.764037 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9"] Mar 20 16:15:19 crc kubenswrapper[4936]: W0320 16:15:19.766817 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4efa574_8274_4c0b_9463_d076e31d2a7e.slice/crio-63513515bcaadd298150aec93934284b61f105a148f1a4fa8d45f804ed61d0db WatchSource:0}: Error finding container 63513515bcaadd298150aec93934284b61f105a148f1a4fa8d45f804ed61d0db: Status 404 returned error can't find the container with id 63513515bcaadd298150aec93934284b61f105a148f1a4fa8d45f804ed61d0db Mar 20 16:15:19 crc kubenswrapper[4936]: I0320 16:15:19.898428 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" event={"ID":"c4efa574-8274-4c0b-9463-d076e31d2a7e","Type":"ContainerStarted","Data":"63513515bcaadd298150aec93934284b61f105a148f1a4fa8d45f804ed61d0db"} Mar 20 16:15:23 crc kubenswrapper[4936]: I0320 16:15:23.944452 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" event={"ID":"c4efa574-8274-4c0b-9463-d076e31d2a7e","Type":"ContainerStarted","Data":"4f9f7fc9700afee0b193325a669ea4b62a02a6e2bc5c29733f91e1a70e5df24b"} Mar 20 16:15:23 crc kubenswrapper[4936]: I0320 16:15:23.945149 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" Mar 20 16:15:23 crc kubenswrapper[4936]: I0320 16:15:23.970814 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" podStartSLOduration=2.38626622 podStartE2EDuration="5.970793045s" podCreationTimestamp="2026-03-20 16:15:18 +0000 UTC" firstStartedPulling="2026-03-20 16:15:19.76881528 +0000 UTC m=+870.715183095" lastFinishedPulling="2026-03-20 16:15:23.353342105 +0000 UTC m=+874.299709920" observedRunningTime="2026-03-20 16:15:23.968069595 +0000 UTC m=+874.914437420" watchObservedRunningTime="2026-03-20 16:15:23.970793045 +0000 UTC m=+874.917160860" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.302807 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d5d6d"] Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.304168 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.315030 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d5d6d"] Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.408121 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vb9f\" (UniqueName: \"kubernetes.io/projected/6f85a08c-441d-436a-b0d0-26a1634d0d8e-kube-api-access-6vb9f\") pod \"community-operators-d5d6d\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.408194 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-catalog-content\") pod \"community-operators-d5d6d\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.408217 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-utilities\") pod \"community-operators-d5d6d\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.509910 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vb9f\" (UniqueName: \"kubernetes.io/projected/6f85a08c-441d-436a-b0d0-26a1634d0d8e-kube-api-access-6vb9f\") pod \"community-operators-d5d6d\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.509976 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-catalog-content\") pod \"community-operators-d5d6d\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.509996 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-utilities\") pod \"community-operators-d5d6d\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.510576 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-utilities\") pod \"community-operators-d5d6d\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.510622 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-catalog-content\") pod \"community-operators-d5d6d\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.538085 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vb9f\" (UniqueName: \"kubernetes.io/projected/6f85a08c-441d-436a-b0d0-26a1634d0d8e-kube-api-access-6vb9f\") pod \"community-operators-d5d6d\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.625323 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.883043 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d5d6d"] Mar 20 16:15:26 crc kubenswrapper[4936]: I0320 16:15:26.963585 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5d6d" event={"ID":"6f85a08c-441d-436a-b0d0-26a1634d0d8e","Type":"ContainerStarted","Data":"dd55101a3505715cfd1f27bebd71001f22a86fe6cc349dc2b1fc35be03d74ad7"} Mar 20 16:15:27 crc kubenswrapper[4936]: I0320 16:15:27.972522 4936 generic.go:334] "Generic (PLEG): container finished" podID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerID="5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b" exitCode=0 Mar 20 16:15:27 crc kubenswrapper[4936]: I0320 16:15:27.972619 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5d6d" event={"ID":"6f85a08c-441d-436a-b0d0-26a1634d0d8e","Type":"ContainerDied","Data":"5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b"} Mar 20 16:15:28 crc kubenswrapper[4936]: I0320 16:15:28.854947 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:15:28 crc kubenswrapper[4936]: I0320 16:15:28.855015 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:15:28 crc kubenswrapper[4936]: I0320 16:15:28.980346 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5d6d" event={"ID":"6f85a08c-441d-436a-b0d0-26a1634d0d8e","Type":"ContainerStarted","Data":"e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543"} Mar 20 16:15:29 crc kubenswrapper[4936]: I0320 16:15:29.333193 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-658f6bdff6-vrsv9" Mar 20 16:15:29 crc kubenswrapper[4936]: I0320 16:15:29.988904 4936 generic.go:334] "Generic (PLEG): container finished" podID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerID="e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543" exitCode=0 Mar 20 16:15:29 crc kubenswrapper[4936]: I0320 16:15:29.988952 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5d6d" event={"ID":"6f85a08c-441d-436a-b0d0-26a1634d0d8e","Type":"ContainerDied","Data":"e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543"} Mar 20 16:15:30 crc kubenswrapper[4936]: I0320 16:15:30.902947 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5sgkl"] Mar 20 16:15:30 crc kubenswrapper[4936]: I0320 16:15:30.904935 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:30 crc kubenswrapper[4936]: I0320 16:15:30.916759 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5sgkl"] Mar 20 16:15:30 crc kubenswrapper[4936]: I0320 16:15:30.965762 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz6sg\" (UniqueName: \"kubernetes.io/projected/bb634055-7548-4a25-bfdd-efc2e2d09b0a-kube-api-access-nz6sg\") pod \"certified-operators-5sgkl\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:30 crc kubenswrapper[4936]: I0320 16:15:30.965835 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-catalog-content\") pod \"certified-operators-5sgkl\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:30 crc kubenswrapper[4936]: I0320 16:15:30.965862 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-utilities\") pod \"certified-operators-5sgkl\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:30 crc kubenswrapper[4936]: I0320 16:15:30.997758 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5d6d" event={"ID":"6f85a08c-441d-436a-b0d0-26a1634d0d8e","Type":"ContainerStarted","Data":"fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726"} Mar 20 16:15:31 crc kubenswrapper[4936]: I0320 16:15:31.019497 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d5d6d" podStartSLOduration=2.780069042 podStartE2EDuration="5.019479026s" podCreationTimestamp="2026-03-20 16:15:26 +0000 UTC" firstStartedPulling="2026-03-20 16:15:27.975199546 +0000 UTC m=+878.921567371" lastFinishedPulling="2026-03-20 16:15:30.21460954 +0000 UTC m=+881.160977355" observedRunningTime="2026-03-20 16:15:31.016704745 +0000 UTC m=+881.963072580" watchObservedRunningTime="2026-03-20 16:15:31.019479026 +0000 UTC m=+881.965846851" Mar 20 16:15:31 crc kubenswrapper[4936]: I0320 16:15:31.067252 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz6sg\" (UniqueName: \"kubernetes.io/projected/bb634055-7548-4a25-bfdd-efc2e2d09b0a-kube-api-access-nz6sg\") pod \"certified-operators-5sgkl\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:31 crc kubenswrapper[4936]: I0320 16:15:31.067322 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-catalog-content\") pod \"certified-operators-5sgkl\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:31 crc kubenswrapper[4936]: I0320 16:15:31.067346 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-utilities\") pod \"certified-operators-5sgkl\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:31 crc kubenswrapper[4936]: I0320 16:15:31.068266 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-catalog-content\") pod \"certified-operators-5sgkl\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:31 crc kubenswrapper[4936]: I0320 16:15:31.068282 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-utilities\") pod \"certified-operators-5sgkl\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:31 crc kubenswrapper[4936]: I0320 16:15:31.089156 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz6sg\" (UniqueName: \"kubernetes.io/projected/bb634055-7548-4a25-bfdd-efc2e2d09b0a-kube-api-access-nz6sg\") pod \"certified-operators-5sgkl\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:31 crc kubenswrapper[4936]: I0320 16:15:31.224771 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:31 crc kubenswrapper[4936]: I0320 16:15:31.600651 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5sgkl"] Mar 20 16:15:31 crc kubenswrapper[4936]: W0320 16:15:31.603288 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb634055_7548_4a25_bfdd_efc2e2d09b0a.slice/crio-43d661349c64d2fe3f12440297af78347cf03625332389439e62f8dd093aa11e WatchSource:0}: Error finding container 43d661349c64d2fe3f12440297af78347cf03625332389439e62f8dd093aa11e: Status 404 returned error can't find the container with id 43d661349c64d2fe3f12440297af78347cf03625332389439e62f8dd093aa11e Mar 20 16:15:32 crc kubenswrapper[4936]: I0320 16:15:32.003748 4936 generic.go:334] "Generic (PLEG): container finished" podID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerID="0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5" exitCode=0 Mar 20 16:15:32 crc kubenswrapper[4936]: I0320 16:15:32.003796 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sgkl" event={"ID":"bb634055-7548-4a25-bfdd-efc2e2d09b0a","Type":"ContainerDied","Data":"0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5"} Mar 20 16:15:32 crc kubenswrapper[4936]: I0320 16:15:32.004119 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sgkl" event={"ID":"bb634055-7548-4a25-bfdd-efc2e2d09b0a","Type":"ContainerStarted","Data":"43d661349c64d2fe3f12440297af78347cf03625332389439e62f8dd093aa11e"} Mar 20 16:15:35 crc kubenswrapper[4936]: I0320 16:15:35.022466 4936 generic.go:334] "Generic (PLEG): container finished" podID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerID="5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49" exitCode=0 Mar 20 16:15:35 crc kubenswrapper[4936]: I0320 16:15:35.022520 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sgkl" event={"ID":"bb634055-7548-4a25-bfdd-efc2e2d09b0a","Type":"ContainerDied","Data":"5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49"} Mar 20 16:15:36 crc kubenswrapper[4936]: I0320 16:15:36.030979 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sgkl" event={"ID":"bb634055-7548-4a25-bfdd-efc2e2d09b0a","Type":"ContainerStarted","Data":"4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed"} Mar 20 16:15:36 crc kubenswrapper[4936]: I0320 16:15:36.079322 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5sgkl" podStartSLOduration=2.661864799 podStartE2EDuration="6.079306984s" podCreationTimestamp="2026-03-20 16:15:30 +0000 UTC" firstStartedPulling="2026-03-20 16:15:32.005335072 +0000 UTC m=+882.951702887" lastFinishedPulling="2026-03-20 16:15:35.422777257 +0000 UTC m=+886.369145072" observedRunningTime="2026-03-20 16:15:36.061077829 +0000 UTC m=+887.007445644" watchObservedRunningTime="2026-03-20 16:15:36.079306984 +0000 UTC m=+887.025674799" Mar 20 16:15:36 crc kubenswrapper[4936]: I0320 16:15:36.626356 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:36 crc kubenswrapper[4936]: I0320 16:15:36.626434 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:36 crc kubenswrapper[4936]: I0320 16:15:36.666133 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:37 crc kubenswrapper[4936]: I0320 16:15:37.108480 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:38 crc kubenswrapper[4936]: I0320 16:15:38.698881 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d5d6d"] Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.045320 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d5d6d" podUID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerName="registry-server" containerID="cri-o://fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726" gracePeriod=2 Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.424162 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.499971 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vb9f\" (UniqueName: \"kubernetes.io/projected/6f85a08c-441d-436a-b0d0-26a1634d0d8e-kube-api-access-6vb9f\") pod \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.500092 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-utilities\") pod \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.500168 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-catalog-content\") pod \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\" (UID: \"6f85a08c-441d-436a-b0d0-26a1634d0d8e\") " Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.501268 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-utilities" (OuterVolumeSpecName: "utilities") pod "6f85a08c-441d-436a-b0d0-26a1634d0d8e" (UID: "6f85a08c-441d-436a-b0d0-26a1634d0d8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.511482 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f85a08c-441d-436a-b0d0-26a1634d0d8e-kube-api-access-6vb9f" (OuterVolumeSpecName: "kube-api-access-6vb9f") pod "6f85a08c-441d-436a-b0d0-26a1634d0d8e" (UID: "6f85a08c-441d-436a-b0d0-26a1634d0d8e"). InnerVolumeSpecName "kube-api-access-6vb9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.546520 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f85a08c-441d-436a-b0d0-26a1634d0d8e" (UID: "6f85a08c-441d-436a-b0d0-26a1634d0d8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.601289 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.601335 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vb9f\" (UniqueName: \"kubernetes.io/projected/6f85a08c-441d-436a-b0d0-26a1634d0d8e-kube-api-access-6vb9f\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:39 crc kubenswrapper[4936]: I0320 16:15:39.601350 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f85a08c-441d-436a-b0d0-26a1634d0d8e-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.057439 4936 generic.go:334] "Generic (PLEG): container finished" podID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerID="fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726" exitCode=0 Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.057498 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5d6d" event={"ID":"6f85a08c-441d-436a-b0d0-26a1634d0d8e","Type":"ContainerDied","Data":"fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726"} Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.057528 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5d6d" event={"ID":"6f85a08c-441d-436a-b0d0-26a1634d0d8e","Type":"ContainerDied","Data":"dd55101a3505715cfd1f27bebd71001f22a86fe6cc349dc2b1fc35be03d74ad7"} Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.057574 4936 scope.go:117] "RemoveContainer" containerID="fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.057768 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5d6d" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.079184 4936 scope.go:117] "RemoveContainer" containerID="e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.102563 4936 scope.go:117] "RemoveContainer" containerID="5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.102695 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d5d6d"] Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.108239 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d5d6d"] Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.130001 4936 scope.go:117] "RemoveContainer" containerID="fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726" Mar 20 16:15:40 crc kubenswrapper[4936]: E0320 16:15:40.130492 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726\": container with ID starting with fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726 not found: ID does not exist" containerID="fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.130524 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726"} err="failed to get container status \"fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726\": rpc error: code = NotFound desc = could not find container \"fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726\": container with ID starting with fc3c17478476a43159ac53f694249e97945b251bf5af8b82e08233cad46b9726 not found: ID does not exist" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.130557 4936 scope.go:117] "RemoveContainer" containerID="e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543" Mar 20 16:15:40 crc kubenswrapper[4936]: E0320 16:15:40.130870 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543\": container with ID starting with e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543 not found: ID does not exist" containerID="e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.130906 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543"} err="failed to get container status \"e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543\": rpc error: code = NotFound desc = could not find container \"e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543\": container with ID starting with e896085c883fa34f5de8450c00e4711efe639ad98723dd0ae414e3818ce0a543 not found: ID does not exist" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.130930 4936 scope.go:117] "RemoveContainer" containerID="5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b" Mar 20 16:15:40 crc kubenswrapper[4936]: E0320 16:15:40.131185 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b\": container with ID starting with 5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b not found: ID does not exist" containerID="5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b" Mar 20 16:15:40 crc kubenswrapper[4936]: I0320 16:15:40.131212 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b"} err="failed to get container status \"5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b\": rpc error: code = NotFound desc = could not find container \"5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b\": container with ID starting with 5f4b9d83b1380b2426af2ffc6d93cd11ec46f75c1cf23303f5ffe132dfd4d87b not found: ID does not exist" Mar 20 16:15:41 crc kubenswrapper[4936]: I0320 16:15:41.225857 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:41 crc kubenswrapper[4936]: I0320 16:15:41.225914 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:41 crc kubenswrapper[4936]: I0320 16:15:41.271555 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:41 crc kubenswrapper[4936]: I0320 16:15:41.862665 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" path="/var/lib/kubelet/pods/6f85a08c-441d-436a-b0d0-26a1634d0d8e/volumes" Mar 20 16:15:42 crc kubenswrapper[4936]: I0320 16:15:42.132862 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:42 crc kubenswrapper[4936]: I0320 16:15:42.894260 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5sgkl"] Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.088102 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5sgkl" podUID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerName="registry-server" containerID="cri-o://4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed" gracePeriod=2 Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.451513 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.565980 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-utilities\") pod \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.566083 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-catalog-content\") pod \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.566179 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz6sg\" (UniqueName: \"kubernetes.io/projected/bb634055-7548-4a25-bfdd-efc2e2d09b0a-kube-api-access-nz6sg\") pod \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\" (UID: \"bb634055-7548-4a25-bfdd-efc2e2d09b0a\") " Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.567266 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-utilities" (OuterVolumeSpecName: "utilities") pod "bb634055-7548-4a25-bfdd-efc2e2d09b0a" (UID: "bb634055-7548-4a25-bfdd-efc2e2d09b0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.571067 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb634055-7548-4a25-bfdd-efc2e2d09b0a-kube-api-access-nz6sg" (OuterVolumeSpecName: "kube-api-access-nz6sg") pod "bb634055-7548-4a25-bfdd-efc2e2d09b0a" (UID: "bb634055-7548-4a25-bfdd-efc2e2d09b0a"). InnerVolumeSpecName "kube-api-access-nz6sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.620880 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb634055-7548-4a25-bfdd-efc2e2d09b0a" (UID: "bb634055-7548-4a25-bfdd-efc2e2d09b0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.667403 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz6sg\" (UniqueName: \"kubernetes.io/projected/bb634055-7548-4a25-bfdd-efc2e2d09b0a-kube-api-access-nz6sg\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.667444 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:44 crc kubenswrapper[4936]: I0320 16:15:44.667458 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb634055-7548-4a25-bfdd-efc2e2d09b0a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.096081 4936 generic.go:334] "Generic (PLEG): container finished" podID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerID="4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed" exitCode=0 Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.096127 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sgkl" event={"ID":"bb634055-7548-4a25-bfdd-efc2e2d09b0a","Type":"ContainerDied","Data":"4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed"} Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.096442 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sgkl" event={"ID":"bb634055-7548-4a25-bfdd-efc2e2d09b0a","Type":"ContainerDied","Data":"43d661349c64d2fe3f12440297af78347cf03625332389439e62f8dd093aa11e"} Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.096466 4936 scope.go:117] "RemoveContainer" containerID="4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.096142 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5sgkl" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.125391 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5sgkl"] Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.127784 4936 scope.go:117] "RemoveContainer" containerID="5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.133127 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5sgkl"] Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.154367 4936 scope.go:117] "RemoveContainer" containerID="0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.170831 4936 scope.go:117] "RemoveContainer" containerID="4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed" Mar 20 16:15:45 crc kubenswrapper[4936]: E0320 16:15:45.172767 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed\": container with ID starting with 4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed not found: ID does not exist" containerID="4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.172834 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed"} err="failed to get container status \"4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed\": rpc error: code = NotFound desc = could not find container \"4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed\": container with ID starting with 4376d4df60760cd3cfc5241bba89abd1dfee7219e64ac594b3268a5af3a31eed not found: ID does not exist" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.172863 4936 scope.go:117] "RemoveContainer" containerID="5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49" Mar 20 16:15:45 crc kubenswrapper[4936]: E0320 16:15:45.173352 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49\": container with ID starting with 5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49 not found: ID does not exist" containerID="5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.173406 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49"} err="failed to get container status \"5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49\": rpc error: code = NotFound desc = could not find container \"5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49\": container with ID starting with 5cf1372c027610bba82a3c4872d69c34b2fd03c4765fb42b7eee40c5c6a39f49 not found: ID does not exist" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.173433 4936 scope.go:117] "RemoveContainer" containerID="0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5" Mar 20 16:15:45 crc kubenswrapper[4936]: E0320 16:15:45.173952 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5\": container with ID starting with 0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5 not found: ID does not exist" containerID="0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.173983 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5"} err="failed to get container status \"0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5\": rpc error: code = NotFound desc = could not find container \"0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5\": container with ID starting with 0e5db801df2fc9df627fcc55204c9c47e8cf60a63944224a8b4d2e4787842ca5 not found: ID does not exist" Mar 20 16:15:45 crc kubenswrapper[4936]: I0320 16:15:45.861472 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" path="/var/lib/kubelet/pods/bb634055-7548-4a25-bfdd-efc2e2d09b0a/volumes" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.965849 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7"] Mar 20 16:15:46 crc kubenswrapper[4936]: E0320 16:15:46.966564 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerName="extract-utilities" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.966585 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerName="extract-utilities" Mar 20 16:15:46 crc kubenswrapper[4936]: E0320 16:15:46.966613 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerName="registry-server" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.966626 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerName="registry-server" Mar 20 16:15:46 crc kubenswrapper[4936]: E0320 16:15:46.966642 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerName="extract-utilities" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.966653 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerName="extract-utilities" Mar 20 16:15:46 crc kubenswrapper[4936]: E0320 16:15:46.966669 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerName="extract-content" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.966679 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerName="extract-content" Mar 20 16:15:46 crc kubenswrapper[4936]: E0320 16:15:46.966703 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerName="extract-content" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.966715 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerName="extract-content" Mar 20 16:15:46 crc kubenswrapper[4936]: E0320 16:15:46.966728 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerName="registry-server" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.966738 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerName="registry-server" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.966909 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f85a08c-441d-436a-b0d0-26a1634d0d8e" containerName="registry-server" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.966934 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb634055-7548-4a25-bfdd-efc2e2d09b0a" containerName="registry-server" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.967584 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.968982 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-zbl2s" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.986156 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf"] Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.987110 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.988996 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-mvnvc" Mar 20 16:15:46 crc kubenswrapper[4936]: I0320 16:15:46.992840 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.017613 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.018666 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.021004 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-rrsmt" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.029642 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.034101 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.039249 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.040138 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.042819 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-4vnvc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.048010 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.048730 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.050513 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-chx5p" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.063576 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.075418 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.096083 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stg2b\" (UniqueName: \"kubernetes.io/projected/f7dbf249-277f-4cb7-8148-7b6062cc5414-kube-api-access-stg2b\") pod \"heat-operator-controller-manager-67dd5f86f5-fxfnf\" (UID: \"f7dbf249-277f-4cb7-8148-7b6062cc5414\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.096151 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl6tf\" (UniqueName: \"kubernetes.io/projected/4cb28ddc-4011-42fb-b8c9-05a2e358330a-kube-api-access-cl6tf\") pod \"barbican-operator-controller-manager-59bc569d95-4pkh7\" (UID: \"4cb28ddc-4011-42fb-b8c9-05a2e358330a\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.096181 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj772\" (UniqueName: \"kubernetes.io/projected/aeb73df4-0e66-4108-b4a2-9aa41de2fcd0-kube-api-access-wj772\") pod \"designate-operator-controller-manager-588d4d986b-qk46l\" (UID: \"aeb73df4-0e66-4108-b4a2-9aa41de2fcd0\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.096307 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk2wz\" (UniqueName: \"kubernetes.io/projected/516b1d4c-72c7-4af0-81bf-489c2a598f8a-kube-api-access-sk2wz\") pod \"glance-operator-controller-manager-79df6bcc97-t7zlc\" (UID: \"516b1d4c-72c7-4af0-81bf-489c2a598f8a\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.096349 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w7jj\" (UniqueName: \"kubernetes.io/projected/3f99cb10-909b-4f34-9b66-95ce7440af5d-kube-api-access-4w7jj\") pod \"cinder-operator-controller-manager-8d58dc466-674vf\" (UID: \"3f99cb10-909b-4f34-9b66-95ce7440af5d\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.101596 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.102510 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.106304 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mmkt2" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.151374 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.162370 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.178076 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.178642 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.178811 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.181704 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pbsbt" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.181789 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.181931 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-m5swn" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.197492 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbmzk\" (UniqueName: \"kubernetes.io/projected/d8898bff-2619-4d34-8998-36ef5aef1c5b-kube-api-access-pbmzk\") pod \"horizon-operator-controller-manager-8464cc45fb-t7n5g\" (UID: \"d8898bff-2619-4d34-8998-36ef5aef1c5b\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.197534 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjg29\" (UniqueName: \"kubernetes.io/projected/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-kube-api-access-hjg29\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.197606 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stg2b\" (UniqueName: \"kubernetes.io/projected/f7dbf249-277f-4cb7-8148-7b6062cc5414-kube-api-access-stg2b\") pod \"heat-operator-controller-manager-67dd5f86f5-fxfnf\" (UID: \"f7dbf249-277f-4cb7-8148-7b6062cc5414\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.197638 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.197657 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl6tf\" (UniqueName: \"kubernetes.io/projected/4cb28ddc-4011-42fb-b8c9-05a2e358330a-kube-api-access-cl6tf\") pod \"barbican-operator-controller-manager-59bc569d95-4pkh7\" (UID: \"4cb28ddc-4011-42fb-b8c9-05a2e358330a\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.197681 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj772\" (UniqueName: \"kubernetes.io/projected/aeb73df4-0e66-4108-b4a2-9aa41de2fcd0-kube-api-access-wj772\") pod \"designate-operator-controller-manager-588d4d986b-qk46l\" (UID: \"aeb73df4-0e66-4108-b4a2-9aa41de2fcd0\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.197717 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk2wz\" (UniqueName: \"kubernetes.io/projected/516b1d4c-72c7-4af0-81bf-489c2a598f8a-kube-api-access-sk2wz\") pod \"glance-operator-controller-manager-79df6bcc97-t7zlc\" (UID: \"516b1d4c-72c7-4af0-81bf-489c2a598f8a\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.197734 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w7jj\" (UniqueName: \"kubernetes.io/projected/3f99cb10-909b-4f34-9b66-95ce7440af5d-kube-api-access-4w7jj\") pod \"cinder-operator-controller-manager-8d58dc466-674vf\" (UID: \"3f99cb10-909b-4f34-9b66-95ce7440af5d\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.198383 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.233738 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.234926 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.238898 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk2wz\" (UniqueName: \"kubernetes.io/projected/516b1d4c-72c7-4af0-81bf-489c2a598f8a-kube-api-access-sk2wz\") pod \"glance-operator-controller-manager-79df6bcc97-t7zlc\" (UID: \"516b1d4c-72c7-4af0-81bf-489c2a598f8a\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.238944 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj772\" (UniqueName: \"kubernetes.io/projected/aeb73df4-0e66-4108-b4a2-9aa41de2fcd0-kube-api-access-wj772\") pod \"designate-operator-controller-manager-588d4d986b-qk46l\" (UID: \"aeb73df4-0e66-4108-b4a2-9aa41de2fcd0\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.244665 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-22wrn" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.250678 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w7jj\" (UniqueName: \"kubernetes.io/projected/3f99cb10-909b-4f34-9b66-95ce7440af5d-kube-api-access-4w7jj\") pod \"cinder-operator-controller-manager-8d58dc466-674vf\" (UID: \"3f99cb10-909b-4f34-9b66-95ce7440af5d\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.250741 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stg2b\" (UniqueName: \"kubernetes.io/projected/f7dbf249-277f-4cb7-8148-7b6062cc5414-kube-api-access-stg2b\") pod \"heat-operator-controller-manager-67dd5f86f5-fxfnf\" (UID: \"f7dbf249-277f-4cb7-8148-7b6062cc5414\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.254328 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.255003 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl6tf\" (UniqueName: \"kubernetes.io/projected/4cb28ddc-4011-42fb-b8c9-05a2e358330a-kube-api-access-cl6tf\") pod \"barbican-operator-controller-manager-59bc569d95-4pkh7\" (UID: \"4cb28ddc-4011-42fb-b8c9-05a2e358330a\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.264496 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-jdd86"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.267842 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.275011 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-csxcf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.292254 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.301612 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.303110 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6drnc\" (UniqueName: \"kubernetes.io/projected/9f92f032-d633-4845-8d38-5cbe5bd39313-kube-api-access-6drnc\") pod \"manila-operator-controller-manager-55f864c847-jdd86\" (UID: \"9f92f032-d633-4845-8d38-5cbe5bd39313\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.303153 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbmzk\" (UniqueName: \"kubernetes.io/projected/d8898bff-2619-4d34-8998-36ef5aef1c5b-kube-api-access-pbmzk\") pod \"horizon-operator-controller-manager-8464cc45fb-t7n5g\" (UID: \"d8898bff-2619-4d34-8998-36ef5aef1c5b\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.303184 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjg29\" (UniqueName: \"kubernetes.io/projected/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-kube-api-access-hjg29\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.303209 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdsgz\" (UniqueName: \"kubernetes.io/projected/0fbd2515-7268-4434-b002-f8e44a0e7b0d-kube-api-access-xdsgz\") pod \"keystone-operator-controller-manager-768b96df4c-5mq8g\" (UID: \"0fbd2515-7268-4434-b002-f8e44a0e7b0d\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.303235 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.303256 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtc7q\" (UniqueName: \"kubernetes.io/projected/8f7f9e3c-7ad9-409e-bc82-6c74e3101630-kube-api-access-dtc7q\") pod \"ironic-operator-controller-manager-6f787dddc9-c7cqh\" (UID: \"8f7f9e3c-7ad9-409e-bc82-6c74e3101630\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.304341 4936 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.304394 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert podName:91412ce4-2ae2-4d80-b2d9-1f9825eaeb35 nodeName:}" failed. No retries permitted until 2026-03-20 16:15:47.804377671 +0000 UTC m=+898.750745486 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert") pod "infra-operator-controller-manager-5c7774cdcf-p6qdv" (UID: "91412ce4-2ae2-4d80-b2d9-1f9825eaeb35") : secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.314634 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-jdd86"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.314971 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.334562 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.335839 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.335914 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.353379 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjg29\" (UniqueName: \"kubernetes.io/projected/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-kube-api-access-hjg29\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.353815 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2nl8t" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.363442 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-f48h2"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.364427 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.365015 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.369230 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.373917 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.374808 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-f48h2"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.382144 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-26t7p" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.383327 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.384300 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.387128 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbmzk\" (UniqueName: \"kubernetes.io/projected/d8898bff-2619-4d34-8998-36ef5aef1c5b-kube-api-access-pbmzk\") pod \"horizon-operator-controller-manager-8464cc45fb-t7n5g\" (UID: \"d8898bff-2619-4d34-8998-36ef5aef1c5b\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.387279 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.395175 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.396782 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-74sq9" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.398069 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-blh7m" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.400351 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.405124 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.405853 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhwwm\" (UniqueName: \"kubernetes.io/projected/7aa0f202-e676-4f55-a20e-64ffc6b90339-kube-api-access-xhwwm\") pod \"neutron-operator-controller-manager-767865f676-f48h2\" (UID: \"7aa0f202-e676-4f55-a20e-64ffc6b90339\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.405893 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6drnc\" (UniqueName: \"kubernetes.io/projected/9f92f032-d633-4845-8d38-5cbe5bd39313-kube-api-access-6drnc\") pod \"manila-operator-controller-manager-55f864c847-jdd86\" (UID: \"9f92f032-d633-4845-8d38-5cbe5bd39313\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.405932 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdsgz\" (UniqueName: \"kubernetes.io/projected/0fbd2515-7268-4434-b002-f8e44a0e7b0d-kube-api-access-xdsgz\") pod \"keystone-operator-controller-manager-768b96df4c-5mq8g\" (UID: \"0fbd2515-7268-4434-b002-f8e44a0e7b0d\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.405973 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtc7q\" (UniqueName: \"kubernetes.io/projected/8f7f9e3c-7ad9-409e-bc82-6c74e3101630-kube-api-access-dtc7q\") pod \"ironic-operator-controller-manager-6f787dddc9-c7cqh\" (UID: \"8f7f9e3c-7ad9-409e-bc82-6c74e3101630\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.405996 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6vrm\" (UniqueName: \"kubernetes.io/projected/80f063a2-e5ad-462b-9a61-06eadd11d477-kube-api-access-n6vrm\") pod \"mariadb-operator-controller-manager-67ccfc9778-hrgvc\" (UID: \"80f063a2-e5ad-462b-9a61-06eadd11d477\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.406032 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chdfq\" (UniqueName: \"kubernetes.io/projected/e9f3d1f0-085a-42b4-92c2-b69e78f69667-kube-api-access-chdfq\") pod \"nova-operator-controller-manager-5d488d59fb-cpjvf\" (UID: \"e9f3d1f0-085a-42b4-92c2-b69e78f69667\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.417517 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-clwss"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.418882 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.419360 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.420712 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.424002 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-755gl" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.424859 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-clwss"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.425652 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-zv8bz" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.425792 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.426873 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.428031 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtc7q\" (UniqueName: \"kubernetes.io/projected/8f7f9e3c-7ad9-409e-bc82-6c74e3101630-kube-api-access-dtc7q\") pod \"ironic-operator-controller-manager-6f787dddc9-c7cqh\" (UID: \"8f7f9e3c-7ad9-409e-bc82-6c74e3101630\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.428165 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.428833 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6drnc\" (UniqueName: \"kubernetes.io/projected/9f92f032-d633-4845-8d38-5cbe5bd39313-kube-api-access-6drnc\") pod \"manila-operator-controller-manager-55f864c847-jdd86\" (UID: \"9f92f032-d633-4845-8d38-5cbe5bd39313\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.429331 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.429958 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdsgz\" (UniqueName: \"kubernetes.io/projected/0fbd2515-7268-4434-b002-f8e44a0e7b0d-kube-api-access-xdsgz\") pod \"keystone-operator-controller-manager-768b96df4c-5mq8g\" (UID: \"0fbd2515-7268-4434-b002-f8e44a0e7b0d\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.434069 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-qknc8" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.455856 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.465936 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.478716 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.479498 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.483240 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-c5kpw" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.491875 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.496805 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.497756 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.505050 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-z4nk7" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.506801 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsx8q\" (UniqueName: \"kubernetes.io/projected/34327669-3184-41ae-b81f-60b97f97a74c-kube-api-access-jsx8q\") pod \"swift-operator-controller-manager-c674c5965-fvbgv\" (UID: \"34327669-3184-41ae-b81f-60b97f97a74c\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.506849 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.506883 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lmdv\" (UniqueName: \"kubernetes.io/projected/830190cb-826c-4da0-a31c-f845de9707a9-kube-api-access-2lmdv\") pod \"placement-operator-controller-manager-5784578c99-9n5lb\" (UID: \"830190cb-826c-4da0-a31c-f845de9707a9\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.506918 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbj2l\" (UniqueName: \"kubernetes.io/projected/a92eb460-57c4-49c7-bd39-2ddb71062d32-kube-api-access-gbj2l\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.506949 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6vrm\" (UniqueName: \"kubernetes.io/projected/80f063a2-e5ad-462b-9a61-06eadd11d477-kube-api-access-n6vrm\") pod \"mariadb-operator-controller-manager-67ccfc9778-hrgvc\" (UID: \"80f063a2-e5ad-462b-9a61-06eadd11d477\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.506974 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkrjm\" (UniqueName: \"kubernetes.io/projected/ae8fec9a-8dd7-481d-a053-9185bdd5aa46-kube-api-access-vkrjm\") pod \"ovn-operator-controller-manager-884679f54-clwss\" (UID: \"ae8fec9a-8dd7-481d-a053-9185bdd5aa46\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.507069 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bncrf\" (UniqueName: \"kubernetes.io/projected/d10fa5e0-be5a-4b79-85d9-a7451ec86255-kube-api-access-bncrf\") pod \"octavia-operator-controller-manager-5b9f45d989-dzfhp\" (UID: \"d10fa5e0-be5a-4b79-85d9-a7451ec86255\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.509410 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.512709 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chdfq\" (UniqueName: \"kubernetes.io/projected/e9f3d1f0-085a-42b4-92c2-b69e78f69667-kube-api-access-chdfq\") pod \"nova-operator-controller-manager-5d488d59fb-cpjvf\" (UID: \"e9f3d1f0-085a-42b4-92c2-b69e78f69667\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.512775 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhwwm\" (UniqueName: \"kubernetes.io/projected/7aa0f202-e676-4f55-a20e-64ffc6b90339-kube-api-access-xhwwm\") pod \"neutron-operator-controller-manager-767865f676-f48h2\" (UID: \"7aa0f202-e676-4f55-a20e-64ffc6b90339\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.530017 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.560640 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6vrm\" (UniqueName: \"kubernetes.io/projected/80f063a2-e5ad-462b-9a61-06eadd11d477-kube-api-access-n6vrm\") pod \"mariadb-operator-controller-manager-67ccfc9778-hrgvc\" (UID: \"80f063a2-e5ad-462b-9a61-06eadd11d477\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.580916 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chdfq\" (UniqueName: \"kubernetes.io/projected/e9f3d1f0-085a-42b4-92c2-b69e78f69667-kube-api-access-chdfq\") pod \"nova-operator-controller-manager-5d488d59fb-cpjvf\" (UID: \"e9f3d1f0-085a-42b4-92c2-b69e78f69667\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.589869 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhwwm\" (UniqueName: \"kubernetes.io/projected/7aa0f202-e676-4f55-a20e-64ffc6b90339-kube-api-access-xhwwm\") pod \"neutron-operator-controller-manager-767865f676-f48h2\" (UID: \"7aa0f202-e676-4f55-a20e-64ffc6b90339\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.605907 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.614123 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bncrf\" (UniqueName: \"kubernetes.io/projected/d10fa5e0-be5a-4b79-85d9-a7451ec86255-kube-api-access-bncrf\") pod \"octavia-operator-controller-manager-5b9f45d989-dzfhp\" (UID: \"d10fa5e0-be5a-4b79-85d9-a7451ec86255\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.614181 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsx8q\" (UniqueName: \"kubernetes.io/projected/34327669-3184-41ae-b81f-60b97f97a74c-kube-api-access-jsx8q\") pod \"swift-operator-controller-manager-c674c5965-fvbgv\" (UID: \"34327669-3184-41ae-b81f-60b97f97a74c\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.614206 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfpr6\" (UniqueName: \"kubernetes.io/projected/7503f97b-163f-40bd-bc93-7d8310c51965-kube-api-access-zfpr6\") pod \"telemetry-operator-controller-manager-d6b694c5-5pfk5\" (UID: \"7503f97b-163f-40bd-bc93-7d8310c51965\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.614404 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.614445 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lmdv\" (UniqueName: \"kubernetes.io/projected/830190cb-826c-4da0-a31c-f845de9707a9-kube-api-access-2lmdv\") pod \"placement-operator-controller-manager-5784578c99-9n5lb\" (UID: \"830190cb-826c-4da0-a31c-f845de9707a9\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.614488 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbj2l\" (UniqueName: \"kubernetes.io/projected/a92eb460-57c4-49c7-bd39-2ddb71062d32-kube-api-access-gbj2l\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.614523 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkrjm\" (UniqueName: \"kubernetes.io/projected/ae8fec9a-8dd7-481d-a053-9185bdd5aa46-kube-api-access-vkrjm\") pod \"ovn-operator-controller-manager-884679f54-clwss\" (UID: \"ae8fec9a-8dd7-481d-a053-9185bdd5aa46\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.614717 4936 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.614852 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert podName:a92eb460-57c4-49c7-bd39-2ddb71062d32 nodeName:}" failed. No retries permitted until 2026-03-20 16:15:48.114826927 +0000 UTC m=+899.061194742 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert") pod "openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" (UID: "a92eb460-57c4-49c7-bd39-2ddb71062d32") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.639993 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.640860 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.644666 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-h77p8" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.646006 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.646283 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbj2l\" (UniqueName: \"kubernetes.io/projected/a92eb460-57c4-49c7-bd39-2ddb71062d32-kube-api-access-gbj2l\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.648285 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsx8q\" (UniqueName: \"kubernetes.io/projected/34327669-3184-41ae-b81f-60b97f97a74c-kube-api-access-jsx8q\") pod \"swift-operator-controller-manager-c674c5965-fvbgv\" (UID: \"34327669-3184-41ae-b81f-60b97f97a74c\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.651223 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lmdv\" (UniqueName: \"kubernetes.io/projected/830190cb-826c-4da0-a31c-f845de9707a9-kube-api-access-2lmdv\") pod \"placement-operator-controller-manager-5784578c99-9n5lb\" (UID: \"830190cb-826c-4da0-a31c-f845de9707a9\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.652496 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bncrf\" (UniqueName: \"kubernetes.io/projected/d10fa5e0-be5a-4b79-85d9-a7451ec86255-kube-api-access-bncrf\") pod \"octavia-operator-controller-manager-5b9f45d989-dzfhp\" (UID: \"d10fa5e0-be5a-4b79-85d9-a7451ec86255\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.656394 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkrjm\" (UniqueName: \"kubernetes.io/projected/ae8fec9a-8dd7-481d-a053-9185bdd5aa46-kube-api-access-vkrjm\") pod \"ovn-operator-controller-manager-884679f54-clwss\" (UID: \"ae8fec9a-8dd7-481d-a053-9185bdd5aa46\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.659277 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.675092 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.679807 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.681710 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.686392 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-p558d" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.686702 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.711849 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.716487 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfpr6\" (UniqueName: \"kubernetes.io/projected/7503f97b-163f-40bd-bc93-7d8310c51965-kube-api-access-zfpr6\") pod \"telemetry-operator-controller-manager-d6b694c5-5pfk5\" (UID: \"7503f97b-163f-40bd-bc93-7d8310c51965\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.716605 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqwx9\" (UniqueName: \"kubernetes.io/projected/1420e79c-a73c-4ed8-902e-eb656a065140-kube-api-access-dqwx9\") pod \"test-operator-controller-manager-5c5cb9c4d7-nbhxk\" (UID: \"1420e79c-a73c-4ed8-902e-eb656a065140\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.716640 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnchq\" (UniqueName: \"kubernetes.io/projected/6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7-kube-api-access-fnchq\") pod \"watcher-operator-controller-manager-6c4d75f7f9-q7pkz\" (UID: \"6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.726076 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.735628 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.737177 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.740347 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.744155 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xp7xn" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.744329 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.744440 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.749508 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfpr6\" (UniqueName: \"kubernetes.io/projected/7503f97b-163f-40bd-bc93-7d8310c51965-kube-api-access-zfpr6\") pod \"telemetry-operator-controller-manager-d6b694c5-5pfk5\" (UID: \"7503f97b-163f-40bd-bc93-7d8310c51965\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.779837 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.787052 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.800566 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.817475 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.817511 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqwx9\" (UniqueName: \"kubernetes.io/projected/1420e79c-a73c-4ed8-902e-eb656a065140-kube-api-access-dqwx9\") pod \"test-operator-controller-manager-5c5cb9c4d7-nbhxk\" (UID: \"1420e79c-a73c-4ed8-902e-eb656a065140\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.817534 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnchq\" (UniqueName: \"kubernetes.io/projected/6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7-kube-api-access-fnchq\") pod \"watcher-operator-controller-manager-6c4d75f7f9-q7pkz\" (UID: \"6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.817584 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.817605 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.817627 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6scp\" (UniqueName: \"kubernetes.io/projected/79279548-dd43-4465-8d51-399bf8ed16ab-kube-api-access-n6scp\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.817647 4936 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.817711 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert podName:91412ce4-2ae2-4d80-b2d9-1f9825eaeb35 nodeName:}" failed. No retries permitted until 2026-03-20 16:15:48.817691257 +0000 UTC m=+899.764059072 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert") pod "infra-operator-controller-manager-5c7774cdcf-p6qdv" (UID: "91412ce4-2ae2-4d80-b2d9-1f9825eaeb35") : secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.835998 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnchq\" (UniqueName: \"kubernetes.io/projected/6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7-kube-api-access-fnchq\") pod \"watcher-operator-controller-manager-6c4d75f7f9-q7pkz\" (UID: \"6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.841467 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqwx9\" (UniqueName: \"kubernetes.io/projected/1420e79c-a73c-4ed8-902e-eb656a065140-kube-api-access-dqwx9\") pod \"test-operator-controller-manager-5c5cb9c4d7-nbhxk\" (UID: \"1420e79c-a73c-4ed8-902e-eb656a065140\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.860156 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.885387 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7"] Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.921902 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.921954 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.921985 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6scp\" (UniqueName: \"kubernetes.io/projected/79279548-dd43-4465-8d51-399bf8ed16ab-kube-api-access-n6scp\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.923474 4936 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.923529 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:15:48.42351001 +0000 UTC m=+899.369877925 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "metrics-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.923863 4936 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: E0320 16:15:47.923901 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:15:48.423890312 +0000 UTC m=+899.370258217 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "webhook-server-cert" not found Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.942246 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6scp\" (UniqueName: \"kubernetes.io/projected/79279548-dd43-4465-8d51-399bf8ed16ab-kube-api-access-n6scp\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:47 crc kubenswrapper[4936]: I0320 16:15:47.976951 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.017395 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.046857 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.135744 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.135880 4936 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.135921 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert podName:a92eb460-57c4-49c7-bd39-2ddb71062d32 nodeName:}" failed. No retries permitted until 2026-03-20 16:15:49.13590795 +0000 UTC m=+900.082275765 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert") pod "openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" (UID: "a92eb460-57c4-49c7-bd39-2ddb71062d32") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.143751 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf"] Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.158477 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" event={"ID":"4cb28ddc-4011-42fb-b8c9-05a2e358330a","Type":"ContainerStarted","Data":"10157a5cdf618c54fe469fbca44329910517d5030e5450fc5425157746166bf6"} Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.218981 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f99cb10_909b_4f34_9b66_95ce7440af5d.slice/crio-c1ed195aad11e2650fa39d4c1372fa01f3f991ee5b44c1bc4af570b85aa0accb WatchSource:0}: Error finding container c1ed195aad11e2650fa39d4c1372fa01f3f991ee5b44c1bc4af570b85aa0accb: Status 404 returned error can't find the container with id c1ed195aad11e2650fa39d4c1372fa01f3f991ee5b44c1bc4af570b85aa0accb Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.436179 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l"] Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.441969 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.442017 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.442112 4936 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.442180 4936 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.442187 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:15:49.442169283 +0000 UTC m=+900.388537098 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "webhook-server-cert" not found Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.442255 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:15:49.442232725 +0000 UTC m=+900.388600660 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "metrics-server-cert" not found Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.442527 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf"] Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.538383 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-jdd86"] Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.547370 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc"] Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.553762 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g"] Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.557504 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod516b1d4c_72c7_4af0_81bf_489c2a598f8a.slice/crio-bd1d4bbe9b3e3582d21e4922f714835ed0a1b2ca1adbb8d3ef7d5f819f6e5d95 WatchSource:0}: Error finding container bd1d4bbe9b3e3582d21e4922f714835ed0a1b2ca1adbb8d3ef7d5f819f6e5d95: Status 404 returned error can't find the container with id bd1d4bbe9b3e3582d21e4922f714835ed0a1b2ca1adbb8d3ef7d5f819f6e5d95 Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.581647 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv"] Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.587678 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fbd2515_7268_4434_b002_f8e44a0e7b0d.slice/crio-f93b2faf12bcc4252106043af728d1bc8ae427f868b0d4d4cefcbd0b5b677327 WatchSource:0}: Error finding container f93b2faf12bcc4252106043af728d1bc8ae427f868b0d4d4cefcbd0b5b677327: Status 404 returned error can't find the container with id f93b2faf12bcc4252106043af728d1bc8ae427f868b0d4d4cefcbd0b5b677327 Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.587965 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g"] Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.601643 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh"] Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.608210 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34327669_3184_41ae_b81f_60b97f97a74c.slice/crio-e399e87258007bd93d6c80a986b2d6fed42921ffdd999f899d1f40a96c8aa0f8 WatchSource:0}: Error finding container e399e87258007bd93d6c80a986b2d6fed42921ffdd999f899d1f40a96c8aa0f8: Status 404 returned error can't find the container with id e399e87258007bd93d6c80a986b2d6fed42921ffdd999f899d1f40a96c8aa0f8 Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.608571 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f7f9e3c_7ad9_409e_bc82_6c74e3101630.slice/crio-73e3f50928d2b61b23ea0689a69cb79d8bf06cdf83d530ba97f9e4e8f63655fd WatchSource:0}: Error finding container 73e3f50928d2b61b23ea0689a69cb79d8bf06cdf83d530ba97f9e4e8f63655fd: Status 404 returned error can't find the container with id 73e3f50928d2b61b23ea0689a69cb79d8bf06cdf83d530ba97f9e4e8f63655fd Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.723766 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-f48h2"] Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.734454 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd10fa5e0_be5a_4b79_85d9_a7451ec86255.slice/crio-427efa1c300b40a4fb23fdcf98ee5aaaf89f05bf80802e25e2d3d33f0b5e8bd6 WatchSource:0}: Error finding container 427efa1c300b40a4fb23fdcf98ee5aaaf89f05bf80802e25e2d3d33f0b5e8bd6: Status 404 returned error can't find the container with id 427efa1c300b40a4fb23fdcf98ee5aaaf89f05bf80802e25e2d3d33f0b5e8bd6 Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.735472 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp"] Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.739078 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7aa0f202_e676_4f55_a20e_64ffc6b90339.slice/crio-c93ae345f41d928fa4b40c6a530006ca81e642b4e654761afbe0bd55fe2bf398 WatchSource:0}: Error finding container c93ae345f41d928fa4b40c6a530006ca81e642b4e654761afbe0bd55fe2bf398: Status 404 returned error can't find the container with id c93ae345f41d928fa4b40c6a530006ca81e642b4e654761afbe0bd55fe2bf398 Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.743958 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb"] Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.750912 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf"] Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.753337 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2lmdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5784578c99-9n5lb_openstack-operators(830190cb-826c-4da0-a31c-f845de9707a9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.754502 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" podUID="830190cb-826c-4da0-a31c-f845de9707a9" Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.756016 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n6vrm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-67ccfc9778-hrgvc_openstack-operators(80f063a2-e5ad-462b-9a61-06eadd11d477): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.764817 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" podUID="80f063a2-e5ad-462b-9a61-06eadd11d477" Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.780489 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc"] Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.846961 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.847095 4936 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.847148 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert podName:91412ce4-2ae2-4d80-b2d9-1f9825eaeb35 nodeName:}" failed. No retries permitted until 2026-03-20 16:15:50.84713497 +0000 UTC m=+901.793502785 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert") pod "infra-operator-controller-manager-5c7774cdcf-p6qdv" (UID: "91412ce4-2ae2-4d80-b2d9-1f9825eaeb35") : secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.911198 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk"] Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.915874 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae8fec9a_8dd7_481d_a053_9185bdd5aa46.slice/crio-ef37c369c759464678952830b73fc4e2ce97ffb967df240a8ab70763c8e687b6 WatchSource:0}: Error finding container ef37c369c759464678952830b73fc4e2ce97ffb967df240a8ab70763c8e687b6: Status 404 returned error can't find the container with id ef37c369c759464678952830b73fc4e2ce97ffb967df240a8ab70763c8e687b6 Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.917961 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1420e79c_a73c_4ed8_902e_eb656a065140.slice/crio-caa635f29e078992191cc74f872baa593339da402e158048bbc8a48ae4205af8 WatchSource:0}: Error finding container caa635f29e078992191cc74f872baa593339da402e158048bbc8a48ae4205af8: Status 404 returned error can't find the container with id caa635f29e078992191cc74f872baa593339da402e158048bbc8a48ae4205af8 Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.919033 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6944a43f_5df3_47b3_ac8c_a6f90eb0cfc7.slice/crio-e9811afa3f4c37dd7d37549d204c2f321b831ea5420d0f1a17110f247dec39a9 WatchSource:0}: Error finding container e9811afa3f4c37dd7d37549d204c2f321b831ea5420d0f1a17110f247dec39a9: Status 404 returned error can't find the container with id e9811afa3f4c37dd7d37549d204c2f321b831ea5420d0f1a17110f247dec39a9 Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.920475 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dqwx9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-nbhxk_openstack-operators(1420e79c-a73c-4ed8-902e-eb656a065140): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.920618 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5"] Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.920846 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fnchq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c4d75f7f9-q7pkz_openstack-operators(6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.921890 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" podUID="1420e79c-a73c-4ed8-902e-eb656a065140" Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.921946 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" podUID="6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7" Mar 20 16:15:48 crc kubenswrapper[4936]: W0320 16:15:48.925277 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7503f97b_163f_40bd_bc93_7d8310c51965.slice/crio-4c2d521d668d500d14845d29295a392b06dab502989b75fc9dbb37cdfc37e5b0 WatchSource:0}: Error finding container 4c2d521d668d500d14845d29295a392b06dab502989b75fc9dbb37cdfc37e5b0: Status 404 returned error can't find the container with id 4c2d521d668d500d14845d29295a392b06dab502989b75fc9dbb37cdfc37e5b0 Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.927487 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zfpr6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d6b694c5-5pfk5_openstack-operators(7503f97b-163f-40bd-bc93-7d8310c51965): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 16:15:48 crc kubenswrapper[4936]: E0320 16:15:48.928792 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" podUID="7503f97b-163f-40bd-bc93-7d8310c51965" Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.931169 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz"] Mar 20 16:15:48 crc kubenswrapper[4936]: I0320 16:15:48.937087 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-clwss"] Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.151626 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.151807 4936 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.151883 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert podName:a92eb460-57c4-49c7-bd39-2ddb71062d32 nodeName:}" failed. No retries permitted until 2026-03-20 16:15:51.151860498 +0000 UTC m=+902.098228313 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert") pod "openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" (UID: "a92eb460-57c4-49c7-bd39-2ddb71062d32") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.166379 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" event={"ID":"1420e79c-a73c-4ed8-902e-eb656a065140","Type":"ContainerStarted","Data":"caa635f29e078992191cc74f872baa593339da402e158048bbc8a48ae4205af8"} Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.168441 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" podUID="1420e79c-a73c-4ed8-902e-eb656a065140" Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.169700 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" event={"ID":"ae8fec9a-8dd7-481d-a053-9185bdd5aa46","Type":"ContainerStarted","Data":"ef37c369c759464678952830b73fc4e2ce97ffb967df240a8ab70763c8e687b6"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.174298 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" event={"ID":"d8898bff-2619-4d34-8998-36ef5aef1c5b","Type":"ContainerStarted","Data":"1d016fb79975140286de2dc542b0a725dffbde5d207bd7fcf27075f37d197811"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.176858 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" event={"ID":"80f063a2-e5ad-462b-9a61-06eadd11d477","Type":"ContainerStarted","Data":"70b3468f32667716a705128e2bbf1fe070a4ddd442f9a6875c29dae04ddc18b4"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.182474 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" event={"ID":"6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7","Type":"ContainerStarted","Data":"e9811afa3f4c37dd7d37549d204c2f321b831ea5420d0f1a17110f247dec39a9"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.184834 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" event={"ID":"aeb73df4-0e66-4108-b4a2-9aa41de2fcd0","Type":"ContainerStarted","Data":"912b6c0ebd5d87f9b392149e52eefcea6541695b83b49ee977717c33225f6e09"} Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.186678 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" podUID="6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7" Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.188956 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" event={"ID":"3f99cb10-909b-4f34-9b66-95ce7440af5d","Type":"ContainerStarted","Data":"c1ed195aad11e2650fa39d4c1372fa01f3f991ee5b44c1bc4af570b85aa0accb"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.190721 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" event={"ID":"0fbd2515-7268-4434-b002-f8e44a0e7b0d","Type":"ContainerStarted","Data":"f93b2faf12bcc4252106043af728d1bc8ae427f868b0d4d4cefcbd0b5b677327"} Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.190930 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" podUID="80f063a2-e5ad-462b-9a61-06eadd11d477" Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.192060 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" event={"ID":"f7dbf249-277f-4cb7-8148-7b6062cc5414","Type":"ContainerStarted","Data":"2c2a288ceffb253805c434637d31aa932c45b9f44d7a0573d4c42e5a41db964b"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.193646 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" event={"ID":"8f7f9e3c-7ad9-409e-bc82-6c74e3101630","Type":"ContainerStarted","Data":"73e3f50928d2b61b23ea0689a69cb79d8bf06cdf83d530ba97f9e4e8f63655fd"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.195501 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" event={"ID":"9f92f032-d633-4845-8d38-5cbe5bd39313","Type":"ContainerStarted","Data":"63b8a93fb23b332700cc7d96b368965b07af1b14610d866cd6b05d90f9638cf5"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.196877 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" event={"ID":"d10fa5e0-be5a-4b79-85d9-a7451ec86255","Type":"ContainerStarted","Data":"427efa1c300b40a4fb23fdcf98ee5aaaf89f05bf80802e25e2d3d33f0b5e8bd6"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.199428 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" event={"ID":"34327669-3184-41ae-b81f-60b97f97a74c","Type":"ContainerStarted","Data":"e399e87258007bd93d6c80a986b2d6fed42921ffdd999f899d1f40a96c8aa0f8"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.204409 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" event={"ID":"516b1d4c-72c7-4af0-81bf-489c2a598f8a","Type":"ContainerStarted","Data":"bd1d4bbe9b3e3582d21e4922f714835ed0a1b2ca1adbb8d3ef7d5f819f6e5d95"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.218417 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" event={"ID":"7503f97b-163f-40bd-bc93-7d8310c51965","Type":"ContainerStarted","Data":"4c2d521d668d500d14845d29295a392b06dab502989b75fc9dbb37cdfc37e5b0"} Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.220117 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" podUID="7503f97b-163f-40bd-bc93-7d8310c51965" Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.221763 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" event={"ID":"e9f3d1f0-085a-42b4-92c2-b69e78f69667","Type":"ContainerStarted","Data":"7824ea702ae297cc0f9bbf67403af85ff55ad64ecfd66b38d3c983a1628a8556"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.223789 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" event={"ID":"7aa0f202-e676-4f55-a20e-64ffc6b90339","Type":"ContainerStarted","Data":"c93ae345f41d928fa4b40c6a530006ca81e642b4e654761afbe0bd55fe2bf398"} Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.226373 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" event={"ID":"830190cb-826c-4da0-a31c-f845de9707a9","Type":"ContainerStarted","Data":"476ff0943b62b702c36e3b1f992bd300bda04727ca2a32e1d15abe7bfbb7fe15"} Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.229602 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" podUID="830190cb-826c-4da0-a31c-f845de9707a9" Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.456952 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:49 crc kubenswrapper[4936]: I0320 16:15:49.457273 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.457146 4936 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.457393 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:15:51.457371949 +0000 UTC m=+902.403739764 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "webhook-server-cert" not found Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.457476 4936 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 16:15:49 crc kubenswrapper[4936]: E0320 16:15:49.457530 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:15:51.457515873 +0000 UTC m=+902.403883688 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "metrics-server-cert" not found Mar 20 16:15:50 crc kubenswrapper[4936]: E0320 16:15:50.239111 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" podUID="830190cb-826c-4da0-a31c-f845de9707a9" Mar 20 16:15:50 crc kubenswrapper[4936]: E0320 16:15:50.239264 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" podUID="6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7" Mar 20 16:15:50 crc kubenswrapper[4936]: E0320 16:15:50.239305 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" podUID="80f063a2-e5ad-462b-9a61-06eadd11d477" Mar 20 16:15:50 crc kubenswrapper[4936]: E0320 16:15:50.239743 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" podUID="1420e79c-a73c-4ed8-902e-eb656a065140" Mar 20 16:15:50 crc kubenswrapper[4936]: E0320 16:15:50.239794 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" podUID="7503f97b-163f-40bd-bc93-7d8310c51965" Mar 20 16:15:50 crc kubenswrapper[4936]: I0320 16:15:50.876118 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:50 crc kubenswrapper[4936]: E0320 16:15:50.876191 4936 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:50 crc kubenswrapper[4936]: E0320 16:15:50.876257 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert podName:91412ce4-2ae2-4d80-b2d9-1f9825eaeb35 nodeName:}" failed. No retries permitted until 2026-03-20 16:15:54.876234325 +0000 UTC m=+905.822602210 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert") pod "infra-operator-controller-manager-5c7774cdcf-p6qdv" (UID: "91412ce4-2ae2-4d80-b2d9-1f9825eaeb35") : secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:51 crc kubenswrapper[4936]: I0320 16:15:51.180155 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:51 crc kubenswrapper[4936]: E0320 16:15:51.180338 4936 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:51 crc kubenswrapper[4936]: E0320 16:15:51.180423 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert podName:a92eb460-57c4-49c7-bd39-2ddb71062d32 nodeName:}" failed. No retries permitted until 2026-03-20 16:15:55.180403047 +0000 UTC m=+906.126770862 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert") pod "openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" (UID: "a92eb460-57c4-49c7-bd39-2ddb71062d32") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:51 crc kubenswrapper[4936]: I0320 16:15:51.484343 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:51 crc kubenswrapper[4936]: I0320 16:15:51.484401 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:51 crc kubenswrapper[4936]: E0320 16:15:51.484505 4936 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 16:15:51 crc kubenswrapper[4936]: E0320 16:15:51.484599 4936 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 16:15:51 crc kubenswrapper[4936]: E0320 16:15:51.484606 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:15:55.484587629 +0000 UTC m=+906.430955444 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "webhook-server-cert" not found Mar 20 16:15:51 crc kubenswrapper[4936]: E0320 16:15:51.484674 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:15:55.48465457 +0000 UTC m=+906.431022435 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "metrics-server-cert" not found Mar 20 16:15:54 crc kubenswrapper[4936]: I0320 16:15:54.956626 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:15:54 crc kubenswrapper[4936]: E0320 16:15:54.956813 4936 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:54 crc kubenswrapper[4936]: E0320 16:15:54.957291 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert podName:91412ce4-2ae2-4d80-b2d9-1f9825eaeb35 nodeName:}" failed. No retries permitted until 2026-03-20 16:16:02.957273904 +0000 UTC m=+913.903641719 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert") pod "infra-operator-controller-manager-5c7774cdcf-p6qdv" (UID: "91412ce4-2ae2-4d80-b2d9-1f9825eaeb35") : secret "infra-operator-webhook-server-cert" not found Mar 20 16:15:55 crc kubenswrapper[4936]: I0320 16:15:55.262289 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:15:55 crc kubenswrapper[4936]: E0320 16:15:55.262668 4936 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:55 crc kubenswrapper[4936]: E0320 16:15:55.262741 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert podName:a92eb460-57c4-49c7-bd39-2ddb71062d32 nodeName:}" failed. No retries permitted until 2026-03-20 16:16:03.262722393 +0000 UTC m=+914.209090208 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert") pod "openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" (UID: "a92eb460-57c4-49c7-bd39-2ddb71062d32") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 16:15:55 crc kubenswrapper[4936]: I0320 16:15:55.565094 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:55 crc kubenswrapper[4936]: I0320 16:15:55.565256 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:15:55 crc kubenswrapper[4936]: E0320 16:15:55.565266 4936 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 16:15:55 crc kubenswrapper[4936]: E0320 16:15:55.565357 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:16:03.565338939 +0000 UTC m=+914.511706754 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "metrics-server-cert" not found Mar 20 16:15:55 crc kubenswrapper[4936]: E0320 16:15:55.565421 4936 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 16:15:55 crc kubenswrapper[4936]: E0320 16:15:55.565518 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:16:03.565494824 +0000 UTC m=+914.511862699 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "webhook-server-cert" not found Mar 20 16:15:58 crc kubenswrapper[4936]: I0320 16:15:58.855338 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:15:58 crc kubenswrapper[4936]: I0320 16:15:58.855750 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.130010 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567056-n6qb7"] Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.131774 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567056-n6qb7" Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.134137 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.134168 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.134253 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.148381 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567056-n6qb7"] Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.241158 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsjqj\" (UniqueName: \"kubernetes.io/projected/bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6-kube-api-access-bsjqj\") pod \"auto-csr-approver-29567056-n6qb7\" (UID: \"bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6\") " pod="openshift-infra/auto-csr-approver-29567056-n6qb7" Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.342455 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsjqj\" (UniqueName: \"kubernetes.io/projected/bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6-kube-api-access-bsjqj\") pod \"auto-csr-approver-29567056-n6qb7\" (UID: \"bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6\") " pod="openshift-infra/auto-csr-approver-29567056-n6qb7" Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.378713 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsjqj\" (UniqueName: \"kubernetes.io/projected/bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6-kube-api-access-bsjqj\") pod \"auto-csr-approver-29567056-n6qb7\" (UID: \"bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6\") " pod="openshift-infra/auto-csr-approver-29567056-n6qb7" Mar 20 16:16:00 crc kubenswrapper[4936]: I0320 16:16:00.473134 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567056-n6qb7" Mar 20 16:16:02 crc kubenswrapper[4936]: I0320 16:16:02.989358 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:16:02 crc kubenswrapper[4936]: I0320 16:16:02.995524 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91412ce4-2ae2-4d80-b2d9-1f9825eaeb35-cert\") pod \"infra-operator-controller-manager-5c7774cdcf-p6qdv\" (UID: \"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35\") " pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:16:03 crc kubenswrapper[4936]: I0320 16:16:03.106855 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pbsbt" Mar 20 16:16:03 crc kubenswrapper[4936]: I0320 16:16:03.115244 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:16:03 crc kubenswrapper[4936]: I0320 16:16:03.293216 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:16:03 crc kubenswrapper[4936]: I0320 16:16:03.299780 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a92eb460-57c4-49c7-bd39-2ddb71062d32-cert\") pod \"openstack-baremetal-operator-controller-manager-86657c54f5gkwd5\" (UID: \"a92eb460-57c4-49c7-bd39-2ddb71062d32\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:16:03 crc kubenswrapper[4936]: I0320 16:16:03.433227 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-zv8bz" Mar 20 16:16:03 crc kubenswrapper[4936]: I0320 16:16:03.442497 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:16:03 crc kubenswrapper[4936]: I0320 16:16:03.597720 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:16:03 crc kubenswrapper[4936]: I0320 16:16:03.597839 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:16:03 crc kubenswrapper[4936]: E0320 16:16:03.597970 4936 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 16:16:03 crc kubenswrapper[4936]: E0320 16:16:03.598076 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs podName:79279548-dd43-4465-8d51-399bf8ed16ab nodeName:}" failed. No retries permitted until 2026-03-20 16:16:19.598052435 +0000 UTC m=+930.544420260 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs") pod "openstack-operator-controller-manager-889c6bbdb-678j6" (UID: "79279548-dd43-4465-8d51-399bf8ed16ab") : secret "webhook-server-cert" not found Mar 20 16:16:03 crc kubenswrapper[4936]: I0320 16:16:03.606105 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-metrics-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:16:09 crc kubenswrapper[4936]: E0320 16:16:09.676999 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a" Mar 20 16:16:09 crc kubenswrapper[4936]: E0320 16:16:09.677795 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-chdfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5d488d59fb-cpjvf_openstack-operators(e9f3d1f0-085a-42b4-92c2-b69e78f69667): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:16:09 crc kubenswrapper[4936]: E0320 16:16:09.678909 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" podUID="e9f3d1f0-085a-42b4-92c2-b69e78f69667" Mar 20 16:16:10 crc kubenswrapper[4936]: E0320 16:16:10.371239 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" podUID="e9f3d1f0-085a-42b4-92c2-b69e78f69667" Mar 20 16:16:10 crc kubenswrapper[4936]: E0320 16:16:10.426513 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56" Mar 20 16:16:10 crc kubenswrapper[4936]: E0320 16:16:10.426725 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xdsgz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-768b96df4c-5mq8g_openstack-operators(0fbd2515-7268-4434-b002-f8e44a0e7b0d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:16:10 crc kubenswrapper[4936]: E0320 16:16:10.428117 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" podUID="0fbd2515-7268-4434-b002-f8e44a0e7b0d" Mar 20 16:16:11 crc kubenswrapper[4936]: E0320 16:16:11.378295 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" podUID="0fbd2515-7268-4434-b002-f8e44a0e7b0d" Mar 20 16:16:13 crc kubenswrapper[4936]: I0320 16:16:13.995092 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567056-n6qb7"] Mar 20 16:16:14 crc kubenswrapper[4936]: W0320 16:16:14.020870 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbe1f0c7_6d84_4883_9ce3_c2d9498c80e6.slice/crio-0e665bb91aa6815d27443824298d4051bcfaa567d48dbb1d38ff6a2a2c3227af WatchSource:0}: Error finding container 0e665bb91aa6815d27443824298d4051bcfaa567d48dbb1d38ff6a2a2c3227af: Status 404 returned error can't find the container with id 0e665bb91aa6815d27443824298d4051bcfaa567d48dbb1d38ff6a2a2c3227af Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.067373 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv"] Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.098217 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5"] Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.394636 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" event={"ID":"a92eb460-57c4-49c7-bd39-2ddb71062d32","Type":"ContainerStarted","Data":"594ae77627c31eec98a461c4501811d347d47b079fe06b1b75c2ca890feb6094"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.399846 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" event={"ID":"3f99cb10-909b-4f34-9b66-95ce7440af5d","Type":"ContainerStarted","Data":"987fe963e17fb7325b585d04c25d6628504d9c9bb0211db6fbd83b298299cdb3"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.399943 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.401836 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" event={"ID":"7aa0f202-e676-4f55-a20e-64ffc6b90339","Type":"ContainerStarted","Data":"ba29dca881e83332776d821ecd7e900a9ab5cb35edd775cc52bd22d905aaddb0"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.401948 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.402890 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" event={"ID":"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35","Type":"ContainerStarted","Data":"820d5e14f61aa1c166c26c50c9fd71146c059eeb0e3b064704a34228ab25f3cf"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.404453 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" event={"ID":"80f063a2-e5ad-462b-9a61-06eadd11d477","Type":"ContainerStarted","Data":"12b78eff1c6ef59f9beed59521d2ee6ca3bfedfaa30f2789c99dcf683f4f2349"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.404601 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.407308 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" event={"ID":"9f92f032-d633-4845-8d38-5cbe5bd39313","Type":"ContainerStarted","Data":"d2f20f9ea93363a49dcdac36636973f0ab02d61f27022eaea7b029d1ecce48da"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.407381 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.414321 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" event={"ID":"34327669-3184-41ae-b81f-60b97f97a74c","Type":"ContainerStarted","Data":"f1766d798628de517231610015f1862317646d2298889746b48deb2d23ff5816"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.414440 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.419732 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" event={"ID":"d8898bff-2619-4d34-8998-36ef5aef1c5b","Type":"ContainerStarted","Data":"2db665885e0ed1cc3a3f9d03499ac008a89c30ed03a4560492a240458159e6a0"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.419802 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.422761 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" podStartSLOduration=6.249682471 podStartE2EDuration="28.422741604s" podCreationTimestamp="2026-03-20 16:15:46 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.222785238 +0000 UTC m=+899.169153053" lastFinishedPulling="2026-03-20 16:16:10.395844371 +0000 UTC m=+921.342212186" observedRunningTime="2026-03-20 16:16:14.420045086 +0000 UTC m=+925.366412901" watchObservedRunningTime="2026-03-20 16:16:14.422741604 +0000 UTC m=+925.369109409" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.424791 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" event={"ID":"7503f97b-163f-40bd-bc93-7d8310c51965","Type":"ContainerStarted","Data":"94e27f7fba6de62d7377b276ac1797f237196789354410beded8b269498f83f4"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.425014 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.436431 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" event={"ID":"8f7f9e3c-7ad9-409e-bc82-6c74e3101630","Type":"ContainerStarted","Data":"0cc711d909316979e4c75f7113cfae1ad73ae51cd6ee46279d544ce26b7a126f"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.436512 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.447183 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" podStartSLOduration=5.083062039 podStartE2EDuration="27.447168156s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.742021767 +0000 UTC m=+899.688389582" lastFinishedPulling="2026-03-20 16:16:11.106127884 +0000 UTC m=+922.052495699" observedRunningTime="2026-03-20 16:16:14.444013156 +0000 UTC m=+925.390380991" watchObservedRunningTime="2026-03-20 16:16:14.447168156 +0000 UTC m=+925.393535971" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.449924 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" event={"ID":"1420e79c-a73c-4ed8-902e-eb656a065140","Type":"ContainerStarted","Data":"21c171e47a342062537e166cedc45d8be42008d6bdde5d8c9e7e2e8b09cedcbe"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.450115 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.454948 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" event={"ID":"6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7","Type":"ContainerStarted","Data":"d703033d7dc7174a7dc946759149df6157c8afa248e073ec5f6fdc609507df12"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.455156 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.456303 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" event={"ID":"830190cb-826c-4da0-a31c-f845de9707a9","Type":"ContainerStarted","Data":"c85a63f941a0034dbbecf68c66786e14aed3d2f4366a4ed94e18610bf4ae5ad7"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.456496 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.457555 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" event={"ID":"aeb73df4-0e66-4108-b4a2-9aa41de2fcd0","Type":"ContainerStarted","Data":"471c6e5bd39a1e0e5b94e8a2d1978f6844a8b2cfaacd3fd28e7ef6f30224fba7"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.457681 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.463557 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" event={"ID":"516b1d4c-72c7-4af0-81bf-489c2a598f8a","Type":"ContainerStarted","Data":"73f035c5bb6e7bf715c883342b6cb236033d2970fa5a8a12353062cf65ca40a6"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.463893 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.468809 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567056-n6qb7" event={"ID":"bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6","Type":"ContainerStarted","Data":"0e665bb91aa6815d27443824298d4051bcfaa567d48dbb1d38ff6a2a2c3227af"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.480679 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" event={"ID":"4cb28ddc-4011-42fb-b8c9-05a2e358330a","Type":"ContainerStarted","Data":"6a895b1819f00b803b8334d9dbcb0b57f4053763ffcb06160cc9015a813a2385"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.480825 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.483993 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" event={"ID":"f7dbf249-277f-4cb7-8148-7b6062cc5414","Type":"ContainerStarted","Data":"b26376a0a7b153bdf9675e54cca28f9621035977bbb775aec61321b6fb16970c"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.484032 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.504328 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" event={"ID":"ae8fec9a-8dd7-481d-a053-9185bdd5aa46","Type":"ContainerStarted","Data":"12274dcca1fa5c2b8dd8c6fafbf96ab2f5fb867232039b10d604ece9ede5ea55"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.504371 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.512938 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" event={"ID":"d10fa5e0-be5a-4b79-85d9-a7451ec86255","Type":"ContainerStarted","Data":"dd92d898dd8fd6bdc13ff3cd09ea7cbfd7adb1f19bfe30b4402d8c028376db9b"} Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.513227 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.539423 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" podStartSLOduration=5.723792001 podStartE2EDuration="27.53940753s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.580346235 +0000 UTC m=+899.526714040" lastFinishedPulling="2026-03-20 16:16:10.395961754 +0000 UTC m=+921.342329569" observedRunningTime="2026-03-20 16:16:14.538434892 +0000 UTC m=+925.484802707" watchObservedRunningTime="2026-03-20 16:16:14.53940753 +0000 UTC m=+925.485775345" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.581214 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" podStartSLOduration=2.767893815 podStartE2EDuration="27.581190892s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.755801311 +0000 UTC m=+899.702169116" lastFinishedPulling="2026-03-20 16:16:13.569098378 +0000 UTC m=+924.515466193" observedRunningTime="2026-03-20 16:16:14.580865243 +0000 UTC m=+925.527233058" watchObservedRunningTime="2026-03-20 16:16:14.581190892 +0000 UTC m=+925.527558707" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.654213 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" podStartSLOduration=5.831599099 podStartE2EDuration="27.654193082s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.610903491 +0000 UTC m=+899.557271306" lastFinishedPulling="2026-03-20 16:16:10.433497474 +0000 UTC m=+921.379865289" observedRunningTime="2026-03-20 16:16:14.639031056 +0000 UTC m=+925.585398871" watchObservedRunningTime="2026-03-20 16:16:14.654193082 +0000 UTC m=+925.600560897" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.765403 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" podStartSLOduration=5.890171555 podStartE2EDuration="27.765380521s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.557172846 +0000 UTC m=+899.503540661" lastFinishedPulling="2026-03-20 16:16:10.432381812 +0000 UTC m=+921.378749627" observedRunningTime="2026-03-20 16:16:14.764410213 +0000 UTC m=+925.710778028" watchObservedRunningTime="2026-03-20 16:16:14.765380521 +0000 UTC m=+925.711748336" Mar 20 16:16:14 crc kubenswrapper[4936]: I0320 16:16:14.957902 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" podStartSLOduration=14.559557544 podStartE2EDuration="27.957882379s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.470972268 +0000 UTC m=+899.417340073" lastFinishedPulling="2026-03-20 16:16:01.869297063 +0000 UTC m=+912.815664908" observedRunningTime="2026-03-20 16:16:14.859704914 +0000 UTC m=+925.806072739" watchObservedRunningTime="2026-03-20 16:16:14.957882379 +0000 UTC m=+925.904250184" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.045556 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" podStartSLOduration=5.588766524 podStartE2EDuration="28.04552256s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.610696865 +0000 UTC m=+899.557064680" lastFinishedPulling="2026-03-20 16:16:11.067452901 +0000 UTC m=+922.013820716" observedRunningTime="2026-03-20 16:16:14.959720761 +0000 UTC m=+925.906088576" watchObservedRunningTime="2026-03-20 16:16:15.04552256 +0000 UTC m=+925.991890375" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.049279 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" podStartSLOduration=3.365565465 podStartE2EDuration="28.049260257s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.927332392 +0000 UTC m=+899.873700207" lastFinishedPulling="2026-03-20 16:16:13.611027184 +0000 UTC m=+924.557394999" observedRunningTime="2026-03-20 16:16:15.04380117 +0000 UTC m=+925.990168985" watchObservedRunningTime="2026-03-20 16:16:15.049260257 +0000 UTC m=+925.995628072" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.087636 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" podStartSLOduration=7.15395991 podStartE2EDuration="29.087614661s" podCreationTimestamp="2026-03-20 16:15:46 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.461925142 +0000 UTC m=+899.408292957" lastFinishedPulling="2026-03-20 16:16:10.395579893 +0000 UTC m=+921.341947708" observedRunningTime="2026-03-20 16:16:15.082251706 +0000 UTC m=+926.028619521" watchObservedRunningTime="2026-03-20 16:16:15.087614661 +0000 UTC m=+926.033982476" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.108809 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" podStartSLOduration=3.455376166 podStartE2EDuration="28.10879043s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.920713468 +0000 UTC m=+899.867081283" lastFinishedPulling="2026-03-20 16:16:13.574127732 +0000 UTC m=+924.520495547" observedRunningTime="2026-03-20 16:16:15.107945796 +0000 UTC m=+926.054313611" watchObservedRunningTime="2026-03-20 16:16:15.10879043 +0000 UTC m=+926.055158255" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.132689 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" podStartSLOduration=5.985667726 podStartE2EDuration="28.132674747s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.920328367 +0000 UTC m=+899.866696182" lastFinishedPulling="2026-03-20 16:16:11.067335388 +0000 UTC m=+922.013703203" observedRunningTime="2026-03-20 16:16:15.128701973 +0000 UTC m=+926.075069798" watchObservedRunningTime="2026-03-20 16:16:15.132674747 +0000 UTC m=+926.079042562" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.242278 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" podStartSLOduration=6.59915188 podStartE2EDuration="28.242260139s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.75234267 +0000 UTC m=+899.698710495" lastFinishedPulling="2026-03-20 16:16:10.395450939 +0000 UTC m=+921.341818754" observedRunningTime="2026-03-20 16:16:15.229535263 +0000 UTC m=+926.175903078" watchObservedRunningTime="2026-03-20 16:16:15.242260139 +0000 UTC m=+926.188627954" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.242503 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" podStartSLOduration=3.572552546 podStartE2EDuration="28.242497626s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.920329947 +0000 UTC m=+899.866697772" lastFinishedPulling="2026-03-20 16:16:13.590275037 +0000 UTC m=+924.536642852" observedRunningTime="2026-03-20 16:16:15.192606681 +0000 UTC m=+926.138974506" watchObservedRunningTime="2026-03-20 16:16:15.242497626 +0000 UTC m=+926.188865441" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.309064 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" podStartSLOduration=3.488157584 podStartE2EDuration="28.309047551s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.753217875 +0000 UTC m=+899.699585690" lastFinishedPulling="2026-03-20 16:16:13.574107842 +0000 UTC m=+924.520475657" observedRunningTime="2026-03-20 16:16:15.307367792 +0000 UTC m=+926.253735617" watchObservedRunningTime="2026-03-20 16:16:15.309047551 +0000 UTC m=+926.255415366" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.309305 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" podStartSLOduration=6.47672624 podStartE2EDuration="28.309281637s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.562260014 +0000 UTC m=+899.508627829" lastFinishedPulling="2026-03-20 16:16:10.394815411 +0000 UTC m=+921.341183226" observedRunningTime="2026-03-20 16:16:15.25689693 +0000 UTC m=+926.203264745" watchObservedRunningTime="2026-03-20 16:16:15.309281637 +0000 UTC m=+926.255649452" Mar 20 16:16:15 crc kubenswrapper[4936]: I0320 16:16:15.362685 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" podStartSLOduration=15.402103568 podStartE2EDuration="29.362669033s" podCreationTimestamp="2026-03-20 16:15:46 +0000 UTC" firstStartedPulling="2026-03-20 16:15:47.964241915 +0000 UTC m=+898.910609730" lastFinishedPulling="2026-03-20 16:16:01.92480736 +0000 UTC m=+912.871175195" observedRunningTime="2026-03-20 16:16:15.360220513 +0000 UTC m=+926.306588328" watchObservedRunningTime="2026-03-20 16:16:15.362669033 +0000 UTC m=+926.309036848" Mar 20 16:16:16 crc kubenswrapper[4936]: I0320 16:16:16.527462 4936 generic.go:334] "Generic (PLEG): container finished" podID="bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6" containerID="f04149f5ca7c1e8786a41e139ac627021de4b5a3a4a6079a996d6a5f0bd416d8" exitCode=0 Mar 20 16:16:16 crc kubenswrapper[4936]: I0320 16:16:16.527584 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567056-n6qb7" event={"ID":"bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6","Type":"ContainerDied","Data":"f04149f5ca7c1e8786a41e139ac627021de4b5a3a4a6079a996d6a5f0bd416d8"} Mar 20 16:16:17 crc kubenswrapper[4936]: I0320 16:16:17.971657 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567056-n6qb7" Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.122654 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsjqj\" (UniqueName: \"kubernetes.io/projected/bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6-kube-api-access-bsjqj\") pod \"bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6\" (UID: \"bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6\") " Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.130928 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6-kube-api-access-bsjqj" (OuterVolumeSpecName: "kube-api-access-bsjqj") pod "bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6" (UID: "bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6"). InnerVolumeSpecName "kube-api-access-bsjqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.223756 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsjqj\" (UniqueName: \"kubernetes.io/projected/bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6-kube-api-access-bsjqj\") on node \"crc\" DevicePath \"\"" Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.541226 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" event={"ID":"91412ce4-2ae2-4d80-b2d9-1f9825eaeb35","Type":"ContainerStarted","Data":"a2a7b19a88267965789ee88c9445181164f2645622b378584805203458106cb4"} Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.541402 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.542805 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" event={"ID":"a92eb460-57c4-49c7-bd39-2ddb71062d32","Type":"ContainerStarted","Data":"f08540580a531908db278e52520f5480196532d19d54ed3eb457dc6ebe9313be"} Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.542895 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.544114 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567056-n6qb7" event={"ID":"bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6","Type":"ContainerDied","Data":"0e665bb91aa6815d27443824298d4051bcfaa567d48dbb1d38ff6a2a2c3227af"} Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.544148 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e665bb91aa6815d27443824298d4051bcfaa567d48dbb1d38ff6a2a2c3227af" Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.544193 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567056-n6qb7" Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.567144 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" podStartSLOduration=27.695266665 podStartE2EDuration="31.567122017s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:16:14.081339683 +0000 UTC m=+925.027707498" lastFinishedPulling="2026-03-20 16:16:17.953195035 +0000 UTC m=+928.899562850" observedRunningTime="2026-03-20 16:16:18.562352679 +0000 UTC m=+929.508720524" watchObservedRunningTime="2026-03-20 16:16:18.567122017 +0000 UTC m=+929.513489882" Mar 20 16:16:18 crc kubenswrapper[4936]: I0320 16:16:18.592590 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" podStartSLOduration=27.744214652 podStartE2EDuration="31.592550908s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:16:14.114766714 +0000 UTC m=+925.061134529" lastFinishedPulling="2026-03-20 16:16:17.96310297 +0000 UTC m=+928.909470785" observedRunningTime="2026-03-20 16:16:18.587093581 +0000 UTC m=+929.533461396" watchObservedRunningTime="2026-03-20 16:16:18.592550908 +0000 UTC m=+929.538918723" Mar 20 16:16:19 crc kubenswrapper[4936]: I0320 16:16:19.023671 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567050-4qjb9"] Mar 20 16:16:19 crc kubenswrapper[4936]: I0320 16:16:19.028326 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567050-4qjb9"] Mar 20 16:16:19 crc kubenswrapper[4936]: I0320 16:16:19.641623 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:16:19 crc kubenswrapper[4936]: I0320 16:16:19.652451 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/79279548-dd43-4465-8d51-399bf8ed16ab-webhook-certs\") pod \"openstack-operator-controller-manager-889c6bbdb-678j6\" (UID: \"79279548-dd43-4465-8d51-399bf8ed16ab\") " pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:16:19 crc kubenswrapper[4936]: I0320 16:16:19.862261 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec3d876a-aacb-4ba2-93cd-9844aa1f70d9" path="/var/lib/kubelet/pods/ec3d876a-aacb-4ba2-93cd-9844aa1f70d9/volumes" Mar 20 16:16:19 crc kubenswrapper[4936]: I0320 16:16:19.934002 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xp7xn" Mar 20 16:16:19 crc kubenswrapper[4936]: I0320 16:16:19.943170 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:16:20 crc kubenswrapper[4936]: I0320 16:16:20.407411 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6"] Mar 20 16:16:20 crc kubenswrapper[4936]: I0320 16:16:20.558645 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" event={"ID":"79279548-dd43-4465-8d51-399bf8ed16ab","Type":"ContainerStarted","Data":"4e0d3ec9b9572c0bbf214c3c3168b0a58b69fc12ec3cd444b945a7ab4baff570"} Mar 20 16:16:21 crc kubenswrapper[4936]: I0320 16:16:21.569195 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" event={"ID":"79279548-dd43-4465-8d51-399bf8ed16ab","Type":"ContainerStarted","Data":"41baf7f15a0f75bb59d8fdee0eb5fabc7e026b8ecf0a8f07d9d1a636fac22311"} Mar 20 16:16:21 crc kubenswrapper[4936]: I0320 16:16:21.569436 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:16:21 crc kubenswrapper[4936]: I0320 16:16:21.606492 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" podStartSLOduration=34.606473591 podStartE2EDuration="34.606473591s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:16:21.601165888 +0000 UTC m=+932.547533703" watchObservedRunningTime="2026-03-20 16:16:21.606473591 +0000 UTC m=+932.552841406" Mar 20 16:16:23 crc kubenswrapper[4936]: I0320 16:16:23.122227 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5c7774cdcf-p6qdv" Mar 20 16:16:23 crc kubenswrapper[4936]: I0320 16:16:23.448099 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-86657c54f5gkwd5" Mar 20 16:16:25 crc kubenswrapper[4936]: I0320 16:16:25.597156 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" event={"ID":"e9f3d1f0-085a-42b4-92c2-b69e78f69667","Type":"ContainerStarted","Data":"b9f00a0060bf1dba3f4b64e9aeda8278d4955c7c8f596cb0ea9108f5b11ca242"} Mar 20 16:16:25 crc kubenswrapper[4936]: I0320 16:16:25.597708 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" Mar 20 16:16:25 crc kubenswrapper[4936]: I0320 16:16:25.613759 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" podStartSLOduration=2.0889689909999998 podStartE2EDuration="38.613742199s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.746968682 +0000 UTC m=+899.693336497" lastFinishedPulling="2026-03-20 16:16:25.27174189 +0000 UTC m=+936.218109705" observedRunningTime="2026-03-20 16:16:25.611085693 +0000 UTC m=+936.557453508" watchObservedRunningTime="2026-03-20 16:16:25.613742199 +0000 UTC m=+936.560110014" Mar 20 16:16:26 crc kubenswrapper[4936]: I0320 16:16:26.607687 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" event={"ID":"0fbd2515-7268-4434-b002-f8e44a0e7b0d","Type":"ContainerStarted","Data":"48ccd5b9210320ea66009f8cb24ae19d2874e8f3ecddd778c89ee2d270079e83"} Mar 20 16:16:26 crc kubenswrapper[4936]: I0320 16:16:26.608260 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" Mar 20 16:16:26 crc kubenswrapper[4936]: I0320 16:16:26.628772 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" podStartSLOduration=1.8481675960000001 podStartE2EDuration="39.628756019s" podCreationTimestamp="2026-03-20 16:15:47 +0000 UTC" firstStartedPulling="2026-03-20 16:15:48.592491931 +0000 UTC m=+899.538859746" lastFinishedPulling="2026-03-20 16:16:26.373080354 +0000 UTC m=+937.319448169" observedRunningTime="2026-03-20 16:16:26.627234275 +0000 UTC m=+937.573602090" watchObservedRunningTime="2026-03-20 16:16:26.628756019 +0000 UTC m=+937.575123834" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.295461 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-4pkh7" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.318893 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-674vf" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.338825 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-qk46l" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.370302 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-t7zlc" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.385088 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-fxfnf" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.429494 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-t7n5g" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.533750 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-c7cqh" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.663007 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-c674c5965-fvbgv" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.678509 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jdd86" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.723392 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hrgvc" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.737479 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-dzfhp" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.781785 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-767865f676-f48h2" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.809496 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-884679f54-clwss" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.866222 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9n5lb" Mar 20 16:16:27 crc kubenswrapper[4936]: I0320 16:16:27.980852 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-5pfk5" Mar 20 16:16:28 crc kubenswrapper[4936]: I0320 16:16:28.019866 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nbhxk" Mar 20 16:16:28 crc kubenswrapper[4936]: I0320 16:16:28.052164 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-q7pkz" Mar 20 16:16:28 crc kubenswrapper[4936]: I0320 16:16:28.854997 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:16:28 crc kubenswrapper[4936]: I0320 16:16:28.855096 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:16:28 crc kubenswrapper[4936]: I0320 16:16:28.855168 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:16:28 crc kubenswrapper[4936]: I0320 16:16:28.856173 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0421fad5e0fd4e5837ad1c312bd061ad68a03b4d2795d6099a681bc4b3ce8f6e"} pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 16:16:28 crc kubenswrapper[4936]: I0320 16:16:28.856255 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" containerID="cri-o://0421fad5e0fd4e5837ad1c312bd061ad68a03b4d2795d6099a681bc4b3ce8f6e" gracePeriod=600 Mar 20 16:16:29 crc kubenswrapper[4936]: I0320 16:16:29.629288 4936 generic.go:334] "Generic (PLEG): container finished" podID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerID="0421fad5e0fd4e5837ad1c312bd061ad68a03b4d2795d6099a681bc4b3ce8f6e" exitCode=0 Mar 20 16:16:29 crc kubenswrapper[4936]: I0320 16:16:29.629368 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerDied","Data":"0421fad5e0fd4e5837ad1c312bd061ad68a03b4d2795d6099a681bc4b3ce8f6e"} Mar 20 16:16:29 crc kubenswrapper[4936]: I0320 16:16:29.629650 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"929e303b68c86406c3ac9c3e153cfe435d16f382342ffc919886659461c4db6e"} Mar 20 16:16:29 crc kubenswrapper[4936]: I0320 16:16:29.629696 4936 scope.go:117] "RemoveContainer" containerID="6e12ccb52ca45cab384d2f9173c7abe5ae747d569fea13ff0f7cd5e5e7c20df0" Mar 20 16:16:29 crc kubenswrapper[4936]: I0320 16:16:29.949256 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-889c6bbdb-678j6" Mar 20 16:16:37 crc kubenswrapper[4936]: I0320 16:16:37.610253 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-5mq8g" Mar 20 16:16:37 crc kubenswrapper[4936]: I0320 16:16:37.790290 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-cpjvf" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.570211 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wpgcx"] Mar 20 16:16:53 crc kubenswrapper[4936]: E0320 16:16:53.576028 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6" containerName="oc" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.576068 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6" containerName="oc" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.576309 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6" containerName="oc" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.578291 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.581571 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.581724 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.581750 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-r8jjj" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.581571 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.583104 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wpgcx"] Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.654754 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cdb5r"] Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.659784 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.662215 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.673522 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cdb5r"] Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.725329 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-config\") pod \"dnsmasq-dns-675f4bcbfc-wpgcx\" (UID: \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.725415 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdbsv\" (UniqueName: \"kubernetes.io/projected/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-kube-api-access-zdbsv\") pod \"dnsmasq-dns-675f4bcbfc-wpgcx\" (UID: \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.725492 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxqtm\" (UniqueName: \"kubernetes.io/projected/8409f27d-2e5c-46f8-8cca-5a678809bf15-kube-api-access-wxqtm\") pod \"dnsmasq-dns-78dd6ddcc-cdb5r\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.725515 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-config\") pod \"dnsmasq-dns-78dd6ddcc-cdb5r\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.725616 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cdb5r\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.827881 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdbsv\" (UniqueName: \"kubernetes.io/projected/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-kube-api-access-zdbsv\") pod \"dnsmasq-dns-675f4bcbfc-wpgcx\" (UID: \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.827984 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxqtm\" (UniqueName: \"kubernetes.io/projected/8409f27d-2e5c-46f8-8cca-5a678809bf15-kube-api-access-wxqtm\") pod \"dnsmasq-dns-78dd6ddcc-cdb5r\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.828013 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-config\") pod \"dnsmasq-dns-78dd6ddcc-cdb5r\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.828060 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cdb5r\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.828121 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-config\") pod \"dnsmasq-dns-675f4bcbfc-wpgcx\" (UID: \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.829000 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cdb5r\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.829201 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-config\") pod \"dnsmasq-dns-78dd6ddcc-cdb5r\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.829231 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-config\") pod \"dnsmasq-dns-675f4bcbfc-wpgcx\" (UID: \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.850325 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdbsv\" (UniqueName: \"kubernetes.io/projected/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-kube-api-access-zdbsv\") pod \"dnsmasq-dns-675f4bcbfc-wpgcx\" (UID: \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.856272 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxqtm\" (UniqueName: \"kubernetes.io/projected/8409f27d-2e5c-46f8-8cca-5a678809bf15-kube-api-access-wxqtm\") pod \"dnsmasq-dns-78dd6ddcc-cdb5r\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.901119 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:16:53 crc kubenswrapper[4936]: I0320 16:16:53.982297 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:16:54 crc kubenswrapper[4936]: I0320 16:16:54.401487 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wpgcx"] Mar 20 16:16:54 crc kubenswrapper[4936]: W0320 16:16:54.408639 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d3cf2aa_b8ef_42ed_9cfd_958b32e7a758.slice/crio-92565d34d6cc0413255b86968f8052329c706887ba64023eef43c532e8401260 WatchSource:0}: Error finding container 92565d34d6cc0413255b86968f8052329c706887ba64023eef43c532e8401260: Status 404 returned error can't find the container with id 92565d34d6cc0413255b86968f8052329c706887ba64023eef43c532e8401260 Mar 20 16:16:54 crc kubenswrapper[4936]: I0320 16:16:54.462794 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cdb5r"] Mar 20 16:16:54 crc kubenswrapper[4936]: W0320 16:16:54.466570 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8409f27d_2e5c_46f8_8cca_5a678809bf15.slice/crio-fafc36dbeabf64343a7b0e38bdde63f363e37a0671a9087780313952dd5c2e3c WatchSource:0}: Error finding container fafc36dbeabf64343a7b0e38bdde63f363e37a0671a9087780313952dd5c2e3c: Status 404 returned error can't find the container with id fafc36dbeabf64343a7b0e38bdde63f363e37a0671a9087780313952dd5c2e3c Mar 20 16:16:54 crc kubenswrapper[4936]: I0320 16:16:54.805266 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" event={"ID":"8409f27d-2e5c-46f8-8cca-5a678809bf15","Type":"ContainerStarted","Data":"fafc36dbeabf64343a7b0e38bdde63f363e37a0671a9087780313952dd5c2e3c"} Mar 20 16:16:54 crc kubenswrapper[4936]: I0320 16:16:54.806602 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" event={"ID":"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758","Type":"ContainerStarted","Data":"92565d34d6cc0413255b86968f8052329c706887ba64023eef43c532e8401260"} Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.255823 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wpgcx"] Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.276348 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mmbhh"] Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.277496 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.302294 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mmbhh"] Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.370635 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fm6x\" (UniqueName: \"kubernetes.io/projected/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-kube-api-access-2fm6x\") pod \"dnsmasq-dns-666b6646f7-mmbhh\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.370708 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-config\") pod \"dnsmasq-dns-666b6646f7-mmbhh\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.370945 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-dns-svc\") pod \"dnsmasq-dns-666b6646f7-mmbhh\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.472120 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fm6x\" (UniqueName: \"kubernetes.io/projected/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-kube-api-access-2fm6x\") pod \"dnsmasq-dns-666b6646f7-mmbhh\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.472196 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-config\") pod \"dnsmasq-dns-666b6646f7-mmbhh\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.472278 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-dns-svc\") pod \"dnsmasq-dns-666b6646f7-mmbhh\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.473428 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-config\") pod \"dnsmasq-dns-666b6646f7-mmbhh\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.473571 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-dns-svc\") pod \"dnsmasq-dns-666b6646f7-mmbhh\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.493110 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fm6x\" (UniqueName: \"kubernetes.io/projected/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-kube-api-access-2fm6x\") pod \"dnsmasq-dns-666b6646f7-mmbhh\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.613283 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.655299 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cdb5r"] Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.698388 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-79vw6"] Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.699744 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.713320 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-79vw6"] Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.780366 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-config\") pod \"dnsmasq-dns-57d769cc4f-79vw6\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.780429 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-79vw6\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.780455 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kxxh\" (UniqueName: \"kubernetes.io/projected/e353d140-f7e5-47b4-b18e-200d5a255086-kube-api-access-8kxxh\") pod \"dnsmasq-dns-57d769cc4f-79vw6\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.882315 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-config\") pod \"dnsmasq-dns-57d769cc4f-79vw6\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.882360 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-79vw6\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.882379 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kxxh\" (UniqueName: \"kubernetes.io/projected/e353d140-f7e5-47b4-b18e-200d5a255086-kube-api-access-8kxxh\") pod \"dnsmasq-dns-57d769cc4f-79vw6\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.886158 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-config\") pod \"dnsmasq-dns-57d769cc4f-79vw6\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.887299 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-79vw6\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:56 crc kubenswrapper[4936]: I0320 16:16:56.903409 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kxxh\" (UniqueName: \"kubernetes.io/projected/e353d140-f7e5-47b4-b18e-200d5a255086-kube-api-access-8kxxh\") pod \"dnsmasq-dns-57d769cc4f-79vw6\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.041247 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:16:57 crc kubenswrapper[4936]: W0320 16:16:57.129534 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce7feb82_b60a_4432_b1b9_b6a11f0e1839.slice/crio-d8318d40d0f80f5decd1ad62ac6313c77c32f02041da5d409897cdd2ceaba73b WatchSource:0}: Error finding container d8318d40d0f80f5decd1ad62ac6313c77c32f02041da5d409897cdd2ceaba73b: Status 404 returned error can't find the container with id d8318d40d0f80f5decd1ad62ac6313c77c32f02041da5d409897cdd2ceaba73b Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.130068 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mmbhh"] Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.222406 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.223960 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.227242 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-q2l4k" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.228047 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.228303 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.228421 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.228536 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.229822 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.229997 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.239387 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.338985 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-79vw6"] Mar 20 16:16:57 crc kubenswrapper[4936]: W0320 16:16:57.342418 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode353d140_f7e5_47b4_b18e_200d5a255086.slice/crio-6de6e379cd2a7715ac771f6329de8a4e268bcd71bade8eab1d4dbb1f2452b1fb WatchSource:0}: Error finding container 6de6e379cd2a7715ac771f6329de8a4e268bcd71bade8eab1d4dbb1f2452b1fb: Status 404 returned error can't find the container with id 6de6e379cd2a7715ac771f6329de8a4e268bcd71bade8eab1d4dbb1f2452b1fb Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.396983 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b3a9af8-80db-4edc-9f6a-42b641c25d71-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397081 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397132 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t5ln\" (UniqueName: \"kubernetes.io/projected/3b3a9af8-80db-4edc-9f6a-42b641c25d71-kube-api-access-9t5ln\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397158 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397251 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b3a9af8-80db-4edc-9f6a-42b641c25d71-config-data\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397309 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b3a9af8-80db-4edc-9f6a-42b641c25d71-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397388 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397422 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b3a9af8-80db-4edc-9f6a-42b641c25d71-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397473 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397517 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b3a9af8-80db-4edc-9f6a-42b641c25d71-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.397566 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.498847 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b3a9af8-80db-4edc-9f6a-42b641c25d71-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.498963 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.498989 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b3a9af8-80db-4edc-9f6a-42b641c25d71-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.499035 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.499064 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b3a9af8-80db-4edc-9f6a-42b641c25d71-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.499123 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.499154 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t5ln\" (UniqueName: \"kubernetes.io/projected/3b3a9af8-80db-4edc-9f6a-42b641c25d71-kube-api-access-9t5ln\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.499194 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.499233 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b3a9af8-80db-4edc-9f6a-42b641c25d71-config-data\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.499603 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.501059 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.501298 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b3a9af8-80db-4edc-9f6a-42b641c25d71-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.501633 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.501707 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b3a9af8-80db-4edc-9f6a-42b641c25d71-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.502223 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b3a9af8-80db-4edc-9f6a-42b641c25d71-config-data\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.502364 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b3a9af8-80db-4edc-9f6a-42b641c25d71-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.502392 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.507271 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.507271 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b3a9af8-80db-4edc-9f6a-42b641c25d71-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.514893 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b3a9af8-80db-4edc-9f6a-42b641c25d71-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.517395 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b3a9af8-80db-4edc-9f6a-42b641c25d71-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.521600 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t5ln\" (UniqueName: \"kubernetes.io/projected/3b3a9af8-80db-4edc-9f6a-42b641c25d71-kube-api-access-9t5ln\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.523329 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"3b3a9af8-80db-4edc-9f6a-42b641c25d71\") " pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.560038 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.614732 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.618151 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.624757 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.625879 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.626056 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.626882 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.626942 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4z4sd" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.627044 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.627151 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.659276 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.711680 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18cfd431-66b8-49d8-8dd9-b2184415c006-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.711757 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.711802 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.711975 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18cfd431-66b8-49d8-8dd9-b2184415c006-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.712068 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4djp\" (UniqueName: \"kubernetes.io/projected/18cfd431-66b8-49d8-8dd9-b2184415c006-kube-api-access-x4djp\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.712121 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.712384 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18cfd431-66b8-49d8-8dd9-b2184415c006-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.712663 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18cfd431-66b8-49d8-8dd9-b2184415c006-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.712691 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.712714 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18cfd431-66b8-49d8-8dd9-b2184415c006-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.712737 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.813685 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.813737 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.813794 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18cfd431-66b8-49d8-8dd9-b2184415c006-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.813814 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4djp\" (UniqueName: \"kubernetes.io/projected/18cfd431-66b8-49d8-8dd9-b2184415c006-kube-api-access-x4djp\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.813848 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.813875 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18cfd431-66b8-49d8-8dd9-b2184415c006-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.813932 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18cfd431-66b8-49d8-8dd9-b2184415c006-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.813959 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.813982 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18cfd431-66b8-49d8-8dd9-b2184415c006-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.814010 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.814028 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18cfd431-66b8-49d8-8dd9-b2184415c006-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.815111 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18cfd431-66b8-49d8-8dd9-b2184415c006-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.815712 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18cfd431-66b8-49d8-8dd9-b2184415c006-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.815981 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.816301 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18cfd431-66b8-49d8-8dd9-b2184415c006-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.816662 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.821300 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18cfd431-66b8-49d8-8dd9-b2184415c006-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.821762 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.828290 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.833204 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18cfd431-66b8-49d8-8dd9-b2184415c006-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.840267 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18cfd431-66b8-49d8-8dd9-b2184415c006-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.847261 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" event={"ID":"e353d140-f7e5-47b4-b18e-200d5a255086","Type":"ContainerStarted","Data":"6de6e379cd2a7715ac771f6329de8a4e268bcd71bade8eab1d4dbb1f2452b1fb"} Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.850214 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" event={"ID":"ce7feb82-b60a-4432-b1b9-b6a11f0e1839","Type":"ContainerStarted","Data":"d8318d40d0f80f5decd1ad62ac6313c77c32f02041da5d409897cdd2ceaba73b"} Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.864757 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4djp\" (UniqueName: \"kubernetes.io/projected/18cfd431-66b8-49d8-8dd9-b2184415c006-kube-api-access-x4djp\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.878305 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18cfd431-66b8-49d8-8dd9-b2184415c006\") " pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:57 crc kubenswrapper[4936]: I0320 16:16:57.965333 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:16:58 crc kubenswrapper[4936]: I0320 16:16:58.010325 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 20 16:16:58 crc kubenswrapper[4936]: I0320 16:16:58.648857 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 20 16:16:58 crc kubenswrapper[4936]: I0320 16:16:58.861059 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18cfd431-66b8-49d8-8dd9-b2184415c006","Type":"ContainerStarted","Data":"cf406ea49244e2caca809ecc43a3dca354cc6f1149af3f4c722ec3757696234f"} Mar 20 16:16:58 crc kubenswrapper[4936]: I0320 16:16:58.863330 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b3a9af8-80db-4edc-9f6a-42b641c25d71","Type":"ContainerStarted","Data":"702618809eadada9d7795dbac262b12083bfe10f8a5f335d5a43f1e04f4ac104"} Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.243818 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.245400 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.248889 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.249509 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-29p7r" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.250874 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.251077 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.264321 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.279075 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.348439 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d8d2157-625a-4b47-a722-1ccbe9236ef6-config-data-default\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.348502 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4swkc\" (UniqueName: \"kubernetes.io/projected/2d8d2157-625a-4b47-a722-1ccbe9236ef6-kube-api-access-4swkc\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.348532 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8d2157-625a-4b47-a722-1ccbe9236ef6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.348582 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d8d2157-625a-4b47-a722-1ccbe9236ef6-kolla-config\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.348668 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.348728 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d8d2157-625a-4b47-a722-1ccbe9236ef6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.349254 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8d2157-625a-4b47-a722-1ccbe9236ef6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.349315 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d8d2157-625a-4b47-a722-1ccbe9236ef6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.450839 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8d2157-625a-4b47-a722-1ccbe9236ef6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.450921 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d8d2157-625a-4b47-a722-1ccbe9236ef6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.450950 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d8d2157-625a-4b47-a722-1ccbe9236ef6-config-data-default\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.450979 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4swkc\" (UniqueName: \"kubernetes.io/projected/2d8d2157-625a-4b47-a722-1ccbe9236ef6-kube-api-access-4swkc\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.450998 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8d2157-625a-4b47-a722-1ccbe9236ef6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.451026 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d8d2157-625a-4b47-a722-1ccbe9236ef6-kolla-config\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.451110 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.451145 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d8d2157-625a-4b47-a722-1ccbe9236ef6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.452599 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.452644 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d8d2157-625a-4b47-a722-1ccbe9236ef6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.453781 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d8d2157-625a-4b47-a722-1ccbe9236ef6-config-data-default\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.453934 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d8d2157-625a-4b47-a722-1ccbe9236ef6-kolla-config\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.455267 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8d2157-625a-4b47-a722-1ccbe9236ef6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.464167 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d8d2157-625a-4b47-a722-1ccbe9236ef6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.467918 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8d2157-625a-4b47-a722-1ccbe9236ef6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.481807 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4swkc\" (UniqueName: \"kubernetes.io/projected/2d8d2157-625a-4b47-a722-1ccbe9236ef6-kube-api-access-4swkc\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.503953 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"2d8d2157-625a-4b47-a722-1ccbe9236ef6\") " pod="openstack/openstack-galera-0" Mar 20 16:16:59 crc kubenswrapper[4936]: I0320 16:16:59.596402 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.294415 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.651344 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.652880 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.655915 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8v4rs" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.656083 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.656254 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.657096 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.685695 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.780487 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82bd48e0-4418-4657-9082-11140724699f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.780596 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/82bd48e0-4418-4657-9082-11140724699f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.780818 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/82bd48e0-4418-4657-9082-11140724699f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.780896 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/82bd48e0-4418-4657-9082-11140724699f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.780917 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.780982 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/82bd48e0-4418-4657-9082-11140724699f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.781054 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwbsg\" (UniqueName: \"kubernetes.io/projected/82bd48e0-4418-4657-9082-11140724699f-kube-api-access-hwbsg\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.781183 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82bd48e0-4418-4657-9082-11140724699f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.882983 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwbsg\" (UniqueName: \"kubernetes.io/projected/82bd48e0-4418-4657-9082-11140724699f-kube-api-access-hwbsg\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.883034 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82bd48e0-4418-4657-9082-11140724699f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.883087 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82bd48e0-4418-4657-9082-11140724699f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.883110 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/82bd48e0-4418-4657-9082-11140724699f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.883145 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/82bd48e0-4418-4657-9082-11140724699f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.883165 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/82bd48e0-4418-4657-9082-11140724699f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.883182 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.883201 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/82bd48e0-4418-4657-9082-11140724699f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.884074 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/82bd48e0-4418-4657-9082-11140724699f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.884115 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/82bd48e0-4418-4657-9082-11140724699f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.885125 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/82bd48e0-4418-4657-9082-11140724699f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.885340 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.886201 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82bd48e0-4418-4657-9082-11140724699f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.947177 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/82bd48e0-4418-4657-9082-11140724699f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.948409 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwbsg\" (UniqueName: \"kubernetes.io/projected/82bd48e0-4418-4657-9082-11140724699f-kube-api-access-hwbsg\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.960739 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82bd48e0-4418-4657-9082-11140724699f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:00 crc kubenswrapper[4936]: I0320 16:17:00.992988 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"82bd48e0-4418-4657-9082-11140724699f\") " pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.226727 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.230427 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.233774 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-vgps6" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.235496 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.235735 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.237654 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.285293 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.302476 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pq8j\" (UniqueName: \"kubernetes.io/projected/6482c360-33e0-4e79-9728-6da33e8adbd1-kube-api-access-5pq8j\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.302533 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6482c360-33e0-4e79-9728-6da33e8adbd1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.302579 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6482c360-33e0-4e79-9728-6da33e8adbd1-kolla-config\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.302680 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6482c360-33e0-4e79-9728-6da33e8adbd1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.302719 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6482c360-33e0-4e79-9728-6da33e8adbd1-config-data\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.404535 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6482c360-33e0-4e79-9728-6da33e8adbd1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.404625 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6482c360-33e0-4e79-9728-6da33e8adbd1-kolla-config\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.404674 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6482c360-33e0-4e79-9728-6da33e8adbd1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.404716 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6482c360-33e0-4e79-9728-6da33e8adbd1-config-data\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.404776 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pq8j\" (UniqueName: \"kubernetes.io/projected/6482c360-33e0-4e79-9728-6da33e8adbd1-kube-api-access-5pq8j\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.405421 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6482c360-33e0-4e79-9728-6da33e8adbd1-kolla-config\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.405655 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6482c360-33e0-4e79-9728-6da33e8adbd1-config-data\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.411940 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6482c360-33e0-4e79-9728-6da33e8adbd1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.425799 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pq8j\" (UniqueName: \"kubernetes.io/projected/6482c360-33e0-4e79-9728-6da33e8adbd1-kube-api-access-5pq8j\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.426283 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6482c360-33e0-4e79-9728-6da33e8adbd1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6482c360-33e0-4e79-9728-6da33e8adbd1\") " pod="openstack/memcached-0" Mar 20 16:17:01 crc kubenswrapper[4936]: I0320 16:17:01.570026 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 20 16:17:03 crc kubenswrapper[4936]: I0320 16:17:03.067054 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 20 16:17:03 crc kubenswrapper[4936]: I0320 16:17:03.071412 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 20 16:17:03 crc kubenswrapper[4936]: I0320 16:17:03.083815 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-bqgn4" Mar 20 16:17:03 crc kubenswrapper[4936]: I0320 16:17:03.092727 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 20 16:17:03 crc kubenswrapper[4936]: I0320 16:17:03.152175 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-747tw\" (UniqueName: \"kubernetes.io/projected/eff64cd4-dfe6-438f-aed8-4c1637cc2db4-kube-api-access-747tw\") pod \"kube-state-metrics-0\" (UID: \"eff64cd4-dfe6-438f-aed8-4c1637cc2db4\") " pod="openstack/kube-state-metrics-0" Mar 20 16:17:03 crc kubenswrapper[4936]: I0320 16:17:03.253595 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-747tw\" (UniqueName: \"kubernetes.io/projected/eff64cd4-dfe6-438f-aed8-4c1637cc2db4-kube-api-access-747tw\") pod \"kube-state-metrics-0\" (UID: \"eff64cd4-dfe6-438f-aed8-4c1637cc2db4\") " pod="openstack/kube-state-metrics-0" Mar 20 16:17:03 crc kubenswrapper[4936]: I0320 16:17:03.280532 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-747tw\" (UniqueName: \"kubernetes.io/projected/eff64cd4-dfe6-438f-aed8-4c1637cc2db4-kube-api-access-747tw\") pod \"kube-state-metrics-0\" (UID: \"eff64cd4-dfe6-438f-aed8-4c1637cc2db4\") " pod="openstack/kube-state-metrics-0" Mar 20 16:17:03 crc kubenswrapper[4936]: I0320 16:17:03.412902 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.404119 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rnftv"] Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.405316 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.406936 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-r9jmr" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.407793 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.413158 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.419330 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rnftv"] Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.456039 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-tp2gk"] Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.461182 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.503631 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-scripts\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.503931 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-var-run\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504054 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-var-log\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504193 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-etc-ovs\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504292 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-var-run\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504390 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q48k\" (UniqueName: \"kubernetes.io/projected/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-kube-api-access-6q48k\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504496 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-var-log-ovn\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504602 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-ovn-controller-tls-certs\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504703 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a660d73e-dab8-4751-8219-c91f6017b574-scripts\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504779 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwvgs\" (UniqueName: \"kubernetes.io/projected/a660d73e-dab8-4751-8219-c91f6017b574-kube-api-access-pwvgs\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504850 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-combined-ca-bundle\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.504915 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-var-lib\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.505008 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-var-run-ovn\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.517953 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-tp2gk"] Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606509 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a660d73e-dab8-4751-8219-c91f6017b574-scripts\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606639 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwvgs\" (UniqueName: \"kubernetes.io/projected/a660d73e-dab8-4751-8219-c91f6017b574-kube-api-access-pwvgs\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606666 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-combined-ca-bundle\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606684 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-var-lib\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606712 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-var-run-ovn\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606733 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-scripts\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606762 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-var-run\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606778 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-var-log\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606812 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-etc-ovs\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606830 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-var-run\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606854 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q48k\" (UniqueName: \"kubernetes.io/projected/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-kube-api-access-6q48k\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606869 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-var-log-ovn\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.606905 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-ovn-controller-tls-certs\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.608191 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-var-log\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.608261 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-var-lib\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.608306 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-var-run-ovn\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.608420 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-var-run\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.608438 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-var-run\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.608703 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-var-log-ovn\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.608758 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a660d73e-dab8-4751-8219-c91f6017b574-etc-ovs\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.609904 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a660d73e-dab8-4751-8219-c91f6017b574-scripts\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.610601 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-scripts\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.615516 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-combined-ca-bundle\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.618198 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-ovn-controller-tls-certs\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.630016 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q48k\" (UniqueName: \"kubernetes.io/projected/cd2037c4-bb05-4f92-9a4a-f71c5da11ebd-kube-api-access-6q48k\") pod \"ovn-controller-rnftv\" (UID: \"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd\") " pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.631576 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwvgs\" (UniqueName: \"kubernetes.io/projected/a660d73e-dab8-4751-8219-c91f6017b574-kube-api-access-pwvgs\") pod \"ovn-controller-ovs-tp2gk\" (UID: \"a660d73e-dab8-4751-8219-c91f6017b574\") " pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.738763 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rnftv" Mar 20 16:17:06 crc kubenswrapper[4936]: I0320 16:17:06.791141 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.028091 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2d8d2157-625a-4b47-a722-1ccbe9236ef6","Type":"ContainerStarted","Data":"4aa6fc3966128e6306d07861ffc0f73afa99533c5e08c1ccb89810b6127e068a"} Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.681855 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.683445 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.685478 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-h2l5s" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.687924 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.690251 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.690407 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.690559 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.690688 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.738502 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.738583 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b0e8ad7-4143-431d-a022-c467ee2d09ac-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.738612 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b0e8ad7-4143-431d-a022-c467ee2d09ac-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.738682 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b0e8ad7-4143-431d-a022-c467ee2d09ac-config\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.738715 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7b0e8ad7-4143-431d-a022-c467ee2d09ac-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.738749 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kkvm\" (UniqueName: \"kubernetes.io/projected/7b0e8ad7-4143-431d-a022-c467ee2d09ac-kube-api-access-9kkvm\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.738772 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b0e8ad7-4143-431d-a022-c467ee2d09ac-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.738797 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0e8ad7-4143-431d-a022-c467ee2d09ac-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.840428 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.840484 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b0e8ad7-4143-431d-a022-c467ee2d09ac-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.840513 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b0e8ad7-4143-431d-a022-c467ee2d09ac-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.840620 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b0e8ad7-4143-431d-a022-c467ee2d09ac-config\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.840766 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.841261 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7b0e8ad7-4143-431d-a022-c467ee2d09ac-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.841502 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b0e8ad7-4143-431d-a022-c467ee2d09ac-config\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.841757 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b0e8ad7-4143-431d-a022-c467ee2d09ac-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.841800 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7b0e8ad7-4143-431d-a022-c467ee2d09ac-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.841866 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kkvm\" (UniqueName: \"kubernetes.io/projected/7b0e8ad7-4143-431d-a022-c467ee2d09ac-kube-api-access-9kkvm\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.842230 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b0e8ad7-4143-431d-a022-c467ee2d09ac-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.842292 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0e8ad7-4143-431d-a022-c467ee2d09ac-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.845290 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b0e8ad7-4143-431d-a022-c467ee2d09ac-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.845440 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b0e8ad7-4143-431d-a022-c467ee2d09ac-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.846876 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0e8ad7-4143-431d-a022-c467ee2d09ac-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.857491 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kkvm\" (UniqueName: \"kubernetes.io/projected/7b0e8ad7-4143-431d-a022-c467ee2d09ac-kube-api-access-9kkvm\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:07 crc kubenswrapper[4936]: I0320 16:17:07.858492 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7b0e8ad7-4143-431d-a022-c467ee2d09ac\") " pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:08 crc kubenswrapper[4936]: I0320 16:17:08.052386 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.078382 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.082360 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.084033 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-f5wvh" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.085113 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.085516 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.085778 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.087887 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.183208 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-config\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.183597 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jxmb\" (UniqueName: \"kubernetes.io/projected/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-kube-api-access-4jxmb\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.183704 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.183731 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.183777 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.183921 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.184073 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.184122 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286039 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286112 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286183 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286218 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286247 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-config\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286268 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jxmb\" (UniqueName: \"kubernetes.io/projected/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-kube-api-access-4jxmb\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286313 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286332 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286351 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.286609 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.288737 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-config\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.289403 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.293274 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.298080 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.301734 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.303968 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jxmb\" (UniqueName: \"kubernetes.io/projected/1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7-kube-api-access-4jxmb\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.308749 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7\") " pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:10 crc kubenswrapper[4936]: I0320 16:17:10.437751 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:13 crc kubenswrapper[4936]: I0320 16:17:13.597125 4936 scope.go:117] "RemoveContainer" containerID="03941498646bcea4ff0136be0cbd242cf6896c3efaada8cca8ad86b0b7e197c6" Mar 20 16:17:21 crc kubenswrapper[4936]: E0320 16:17:21.573775 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 20 16:17:21 crc kubenswrapper[4936]: E0320 16:17:21.574477 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zdbsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-wpgcx_openstack(4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:17:21 crc kubenswrapper[4936]: E0320 16:17:21.575932 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" podUID="4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758" Mar 20 16:17:21 crc kubenswrapper[4936]: E0320 16:17:21.587609 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 20 16:17:21 crc kubenswrapper[4936]: E0320 16:17:21.587790 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2fm6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-mmbhh_openstack(ce7feb82-b60a-4432-b1b9-b6a11f0e1839): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:17:21 crc kubenswrapper[4936]: E0320 16:17:21.589001 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" podUID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" Mar 20 16:17:21 crc kubenswrapper[4936]: E0320 16:17:21.595104 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 20 16:17:21 crc kubenswrapper[4936]: E0320 16:17:21.595285 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8kxxh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-79vw6_openstack(e353d140-f7e5-47b4-b18e-200d5a255086): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:17:21 crc kubenswrapper[4936]: E0320 16:17:21.596468 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" podUID="e353d140-f7e5-47b4-b18e-200d5a255086" Mar 20 16:17:22 crc kubenswrapper[4936]: E0320 16:17:22.138871 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" podUID="e353d140-f7e5-47b4-b18e-200d5a255086" Mar 20 16:17:22 crc kubenswrapper[4936]: E0320 16:17:22.139472 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" podUID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.155129 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" event={"ID":"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758","Type":"ContainerDied","Data":"92565d34d6cc0413255b86968f8052329c706887ba64023eef43c532e8401260"} Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.155435 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92565d34d6cc0413255b86968f8052329c706887ba64023eef43c532e8401260" Mar 20 16:17:24 crc kubenswrapper[4936]: E0320 16:17:24.181286 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 20 16:17:24 crc kubenswrapper[4936]: E0320 16:17:24.181432 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wxqtm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-cdb5r_openstack(8409f27d-2e5c-46f8-8cca-5a678809bf15): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:17:24 crc kubenswrapper[4936]: E0320 16:17:24.182729 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" podUID="8409f27d-2e5c-46f8-8cca-5a678809bf15" Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.195744 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.218618 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdbsv\" (UniqueName: \"kubernetes.io/projected/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-kube-api-access-zdbsv\") pod \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\" (UID: \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\") " Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.219010 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-config\") pod \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\" (UID: \"4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758\") " Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.220087 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-config" (OuterVolumeSpecName: "config") pod "4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758" (UID: "4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.252581 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-kube-api-access-zdbsv" (OuterVolumeSpecName: "kube-api-access-zdbsv") pod "4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758" (UID: "4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758"). InnerVolumeSpecName "kube-api-access-zdbsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.320650 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdbsv\" (UniqueName: \"kubernetes.io/projected/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-kube-api-access-zdbsv\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.320698 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.581636 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 20 16:17:24 crc kubenswrapper[4936]: I0320 16:17:24.643220 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 20 16:17:25 crc kubenswrapper[4936]: I0320 16:17:25.161103 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wpgcx" Mar 20 16:17:25 crc kubenswrapper[4936]: I0320 16:17:25.219021 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wpgcx"] Mar 20 16:17:25 crc kubenswrapper[4936]: I0320 16:17:25.224875 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wpgcx"] Mar 20 16:17:25 crc kubenswrapper[4936]: E0320 16:17:25.849897 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/openstack-rabbitmq:r42p" Mar 20 16:17:25 crc kubenswrapper[4936]: E0320 16:17:25.850088 4936 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 16:17:25 crc kubenswrapper[4936]: init container &Container{Name:setup-container,Image:quay.io/lmiccini/openstack-rabbitmq:r42p,Command:[sh -c],Args:[set -e Mar 20 16:17:25 crc kubenswrapper[4936]: cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie Mar 20 16:17:25 crc kubenswrapper[4936]: chmod 600 /var/lib/rabbitmq/.erlang.cookie Mar 20 16:17:25 crc kubenswrapper[4936]: cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins Mar 20 16:17:25 crc kubenswrapper[4936]: echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf Mar 20 16:17:25 crc kubenswrapper[4936]: sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf Mar 20 16:17:25 crc kubenswrapper[4936]: chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf Mar 20 16:17:25 crc kubenswrapper[4936]: # Allow time for multi-pod clusters to complete peer discovery Mar 20 16:17:25 crc kubenswrapper[4936]: sleep 30],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x4djp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(18cfd431-66b8-49d8-8dd9-b2184415c006): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Mar 20 16:17:25 crc kubenswrapper[4936]: > logger="UnhandledError" Mar 20 16:17:25 crc kubenswrapper[4936]: E0320 16:17:25.851365 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="18cfd431-66b8-49d8-8dd9-b2184415c006" Mar 20 16:17:25 crc kubenswrapper[4936]: I0320 16:17:25.871033 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758" path="/var/lib/kubelet/pods/4d3cf2aa-b8ef-42ed-9cfd-958b32e7a758/volumes" Mar 20 16:17:25 crc kubenswrapper[4936]: E0320 16:17:25.920401 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/openstack-rabbitmq:r42p" Mar 20 16:17:25 crc kubenswrapper[4936]: E0320 16:17:25.920616 4936 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 16:17:25 crc kubenswrapper[4936]: init container &Container{Name:setup-container,Image:quay.io/lmiccini/openstack-rabbitmq:r42p,Command:[sh -c],Args:[set -e Mar 20 16:17:25 crc kubenswrapper[4936]: cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie Mar 20 16:17:25 crc kubenswrapper[4936]: chmod 600 /var/lib/rabbitmq/.erlang.cookie Mar 20 16:17:25 crc kubenswrapper[4936]: cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins Mar 20 16:17:25 crc kubenswrapper[4936]: echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf Mar 20 16:17:25 crc kubenswrapper[4936]: sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf Mar 20 16:17:25 crc kubenswrapper[4936]: chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf Mar 20 16:17:25 crc kubenswrapper[4936]: # Allow time for multi-pod clusters to complete peer discovery Mar 20 16:17:25 crc kubenswrapper[4936]: sleep 30],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9t5ln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(3b3a9af8-80db-4edc-9f6a-42b641c25d71): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Mar 20 16:17:25 crc kubenswrapper[4936]: > logger="UnhandledError" Mar 20 16:17:25 crc kubenswrapper[4936]: E0320 16:17:25.924760 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="3b3a9af8-80db-4edc-9f6a-42b641c25d71" Mar 20 16:17:25 crc kubenswrapper[4936]: I0320 16:17:25.978038 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.149988 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-dns-svc\") pod \"8409f27d-2e5c-46f8-8cca-5a678809bf15\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.150673 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8409f27d-2e5c-46f8-8cca-5a678809bf15" (UID: "8409f27d-2e5c-46f8-8cca-5a678809bf15"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.151213 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxqtm\" (UniqueName: \"kubernetes.io/projected/8409f27d-2e5c-46f8-8cca-5a678809bf15-kube-api-access-wxqtm\") pod \"8409f27d-2e5c-46f8-8cca-5a678809bf15\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.151434 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-config\") pod \"8409f27d-2e5c-46f8-8cca-5a678809bf15\" (UID: \"8409f27d-2e5c-46f8-8cca-5a678809bf15\") " Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.151941 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-config" (OuterVolumeSpecName: "config") pod "8409f27d-2e5c-46f8-8cca-5a678809bf15" (UID: "8409f27d-2e5c-46f8-8cca-5a678809bf15"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.152260 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.152282 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8409f27d-2e5c-46f8-8cca-5a678809bf15-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.155234 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8409f27d-2e5c-46f8-8cca-5a678809bf15-kube-api-access-wxqtm" (OuterVolumeSpecName: "kube-api-access-wxqtm") pod "8409f27d-2e5c-46f8-8cca-5a678809bf15" (UID: "8409f27d-2e5c-46f8-8cca-5a678809bf15"). InnerVolumeSpecName "kube-api-access-wxqtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.171796 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6482c360-33e0-4e79-9728-6da33e8adbd1","Type":"ContainerStarted","Data":"0e597658004d1c75377bea30ba05fa0a58cd81b44deea9b11b670a4810162f43"} Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.176764 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"82bd48e0-4418-4657-9082-11140724699f","Type":"ContainerStarted","Data":"4d1bf8fe0e475a11a0d24b1d56d2164b379bc0d2849becb3a7dc6f072d4cc9d7"} Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.176808 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"82bd48e0-4418-4657-9082-11140724699f","Type":"ContainerStarted","Data":"05634d0fbf3c270e85703865e08636bc358966914b0c843e494dd5474e2da21b"} Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.178780 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2d8d2157-625a-4b47-a722-1ccbe9236ef6","Type":"ContainerStarted","Data":"da36dcd15af113d7cf30f2f9527ee6062a072402f61e623ba0331de96335d2b6"} Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.180983 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" event={"ID":"8409f27d-2e5c-46f8-8cca-5a678809bf15","Type":"ContainerDied","Data":"fafc36dbeabf64343a7b0e38bdde63f363e37a0671a9087780313952dd5c2e3c"} Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.181192 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cdb5r" Mar 20 16:17:26 crc kubenswrapper[4936]: E0320 16:17:26.182674 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/openstack-rabbitmq:r42p\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="18cfd431-66b8-49d8-8dd9-b2184415c006" Mar 20 16:17:26 crc kubenswrapper[4936]: E0320 16:17:26.182834 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/openstack-rabbitmq:r42p\\\"\"" pod="openstack/rabbitmq-server-0" podUID="3b3a9af8-80db-4edc-9f6a-42b641c25d71" Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.254626 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxqtm\" (UniqueName: \"kubernetes.io/projected/8409f27d-2e5c-46f8-8cca-5a678809bf15-kube-api-access-wxqtm\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.332666 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.347124 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cdb5r"] Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.353767 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cdb5r"] Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.404570 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rnftv"] Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.541752 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-tp2gk"] Mar 20 16:17:26 crc kubenswrapper[4936]: W0320 16:17:26.549479 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda660d73e_dab8_4751_8219_c91f6017b574.slice/crio-90070bd336c7d5de9b19a24e588ff3e1136cf3e6c7b013fdb42d56e7f2d88173 WatchSource:0}: Error finding container 90070bd336c7d5de9b19a24e588ff3e1136cf3e6c7b013fdb42d56e7f2d88173: Status 404 returned error can't find the container with id 90070bd336c7d5de9b19a24e588ff3e1136cf3e6c7b013fdb42d56e7f2d88173 Mar 20 16:17:26 crc kubenswrapper[4936]: I0320 16:17:26.629235 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 20 16:17:26 crc kubenswrapper[4936]: W0320 16:17:26.643739 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ba28ba5_07dc_4f0c_ab8a_2d23b6e143e7.slice/crio-2188d4ad9ec4ddfb66d7c59d979217e38c24ffffcea181262212484e239f42d7 WatchSource:0}: Error finding container 2188d4ad9ec4ddfb66d7c59d979217e38c24ffffcea181262212484e239f42d7: Status 404 returned error can't find the container with id 2188d4ad9ec4ddfb66d7c59d979217e38c24ffffcea181262212484e239f42d7 Mar 20 16:17:27 crc kubenswrapper[4936]: I0320 16:17:27.140163 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 20 16:17:27 crc kubenswrapper[4936]: I0320 16:17:27.207178 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7b0e8ad7-4143-431d-a022-c467ee2d09ac","Type":"ContainerStarted","Data":"766f52e76959439481d1f9d5b86a2c15e9f0486ff3b0536fcc18becbbff4aece"} Mar 20 16:17:27 crc kubenswrapper[4936]: I0320 16:17:27.208523 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rnftv" event={"ID":"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd","Type":"ContainerStarted","Data":"5b4eab62c5adbd427655ec7d56f7d7f58874da66ff07503bd78b535f172b1ae2"} Mar 20 16:17:27 crc kubenswrapper[4936]: I0320 16:17:27.216688 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eff64cd4-dfe6-438f-aed8-4c1637cc2db4","Type":"ContainerStarted","Data":"74af9ce3a097763ae1a49769cd4660d866d6f4afbe4834cafcdb215e8a4f02f6"} Mar 20 16:17:27 crc kubenswrapper[4936]: I0320 16:17:27.218604 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7","Type":"ContainerStarted","Data":"2188d4ad9ec4ddfb66d7c59d979217e38c24ffffcea181262212484e239f42d7"} Mar 20 16:17:27 crc kubenswrapper[4936]: I0320 16:17:27.222803 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tp2gk" event={"ID":"a660d73e-dab8-4751-8219-c91f6017b574","Type":"ContainerStarted","Data":"90070bd336c7d5de9b19a24e588ff3e1136cf3e6c7b013fdb42d56e7f2d88173"} Mar 20 16:17:27 crc kubenswrapper[4936]: I0320 16:17:27.865613 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8409f27d-2e5c-46f8-8cca-5a678809bf15" path="/var/lib/kubelet/pods/8409f27d-2e5c-46f8-8cca-5a678809bf15/volumes" Mar 20 16:17:31 crc kubenswrapper[4936]: I0320 16:17:31.266581 4936 generic.go:334] "Generic (PLEG): container finished" podID="82bd48e0-4418-4657-9082-11140724699f" containerID="4d1bf8fe0e475a11a0d24b1d56d2164b379bc0d2849becb3a7dc6f072d4cc9d7" exitCode=0 Mar 20 16:17:31 crc kubenswrapper[4936]: I0320 16:17:31.266744 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"82bd48e0-4418-4657-9082-11140724699f","Type":"ContainerDied","Data":"4d1bf8fe0e475a11a0d24b1d56d2164b379bc0d2849becb3a7dc6f072d4cc9d7"} Mar 20 16:17:31 crc kubenswrapper[4936]: I0320 16:17:31.272276 4936 generic.go:334] "Generic (PLEG): container finished" podID="2d8d2157-625a-4b47-a722-1ccbe9236ef6" containerID="da36dcd15af113d7cf30f2f9527ee6062a072402f61e623ba0331de96335d2b6" exitCode=0 Mar 20 16:17:31 crc kubenswrapper[4936]: I0320 16:17:31.272330 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2d8d2157-625a-4b47-a722-1ccbe9236ef6","Type":"ContainerDied","Data":"da36dcd15af113d7cf30f2f9527ee6062a072402f61e623ba0331de96335d2b6"} Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.293516 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2d8d2157-625a-4b47-a722-1ccbe9236ef6","Type":"ContainerStarted","Data":"1d643d7f9f8c0d52f39a5af169a38bd1cf25152b4b4fed09ea1005ea32bc60fe"} Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.302973 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7b0e8ad7-4143-431d-a022-c467ee2d09ac","Type":"ContainerStarted","Data":"c32f74b23404207bdcc5cf86662c7ecc9917c0153ce792763f2e685791fdd5c7"} Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.306605 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eff64cd4-dfe6-438f-aed8-4c1637cc2db4","Type":"ContainerStarted","Data":"dd91ee26f881be46e6bcd680946f0538c4007735d7ffb3f683b2b971e71ed118"} Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.307387 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.318823 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6482c360-33e0-4e79-9728-6da33e8adbd1","Type":"ContainerStarted","Data":"91f38cf9da0f6a467d0f43f10c4fe0001c69dabf16c5703e6db29b92130b974e"} Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.320011 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.324061 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"82bd48e0-4418-4657-9082-11140724699f","Type":"ContainerStarted","Data":"2e5846114a054438eafc306b46edad22658c642f54d4bcc3d520b35af346213e"} Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.332212 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7","Type":"ContainerStarted","Data":"0abc6df380142863c4e3ffa8092614bee6a53444c618424df003d486634e3ae3"} Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.333720 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tp2gk" event={"ID":"a660d73e-dab8-4751-8219-c91f6017b574","Type":"ContainerStarted","Data":"57dca4e49e04fdc671cfd814d20c48e294e4d597125bee6e95f91472711e7779"} Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.347492 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=16.244832036 podStartE2EDuration="35.347474255s" podCreationTimestamp="2026-03-20 16:16:58 +0000 UTC" firstStartedPulling="2026-03-20 16:17:06.752881634 +0000 UTC m=+977.699249449" lastFinishedPulling="2026-03-20 16:17:25.855523853 +0000 UTC m=+996.801891668" observedRunningTime="2026-03-20 16:17:33.317011537 +0000 UTC m=+1004.263379362" watchObservedRunningTime="2026-03-20 16:17:33.347474255 +0000 UTC m=+1004.293842070" Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.348684 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=23.719862102 podStartE2EDuration="30.348672689s" podCreationTimestamp="2026-03-20 16:17:03 +0000 UTC" firstStartedPulling="2026-03-20 16:17:26.310002828 +0000 UTC m=+997.256370643" lastFinishedPulling="2026-03-20 16:17:32.938813415 +0000 UTC m=+1003.885181230" observedRunningTime="2026-03-20 16:17:33.345975201 +0000 UTC m=+1004.292343016" watchObservedRunningTime="2026-03-20 16:17:33.348672689 +0000 UTC m=+1004.295040514" Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.371300 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=34.371277698 podStartE2EDuration="34.371277698s" podCreationTimestamp="2026-03-20 16:16:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:17:33.368629051 +0000 UTC m=+1004.314996876" watchObservedRunningTime="2026-03-20 16:17:33.371277698 +0000 UTC m=+1004.317645513" Mar 20 16:17:33 crc kubenswrapper[4936]: I0320 16:17:33.415323 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=26.869867565 podStartE2EDuration="32.415285871s" podCreationTimestamp="2026-03-20 16:17:01 +0000 UTC" firstStartedPulling="2026-03-20 16:17:25.874182007 +0000 UTC m=+996.820549822" lastFinishedPulling="2026-03-20 16:17:31.419600313 +0000 UTC m=+1002.365968128" observedRunningTime="2026-03-20 16:17:33.41285065 +0000 UTC m=+1004.359218475" watchObservedRunningTime="2026-03-20 16:17:33.415285871 +0000 UTC m=+1004.361653686" Mar 20 16:17:34 crc kubenswrapper[4936]: I0320 16:17:34.341829 4936 generic.go:334] "Generic (PLEG): container finished" podID="a660d73e-dab8-4751-8219-c91f6017b574" containerID="57dca4e49e04fdc671cfd814d20c48e294e4d597125bee6e95f91472711e7779" exitCode=0 Mar 20 16:17:34 crc kubenswrapper[4936]: I0320 16:17:34.341897 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tp2gk" event={"ID":"a660d73e-dab8-4751-8219-c91f6017b574","Type":"ContainerDied","Data":"57dca4e49e04fdc671cfd814d20c48e294e4d597125bee6e95f91472711e7779"} Mar 20 16:17:34 crc kubenswrapper[4936]: I0320 16:17:34.346016 4936 generic.go:334] "Generic (PLEG): container finished" podID="e353d140-f7e5-47b4-b18e-200d5a255086" containerID="002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c" exitCode=0 Mar 20 16:17:34 crc kubenswrapper[4936]: I0320 16:17:34.348135 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" event={"ID":"e353d140-f7e5-47b4-b18e-200d5a255086","Type":"ContainerDied","Data":"002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c"} Mar 20 16:17:34 crc kubenswrapper[4936]: I0320 16:17:34.348667 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rnftv" event={"ID":"cd2037c4-bb05-4f92-9a4a-f71c5da11ebd","Type":"ContainerStarted","Data":"02a48fb187a605d81cfc97b6f43dc059fb04152929ea6feb84281b585b593e66"} Mar 20 16:17:34 crc kubenswrapper[4936]: I0320 16:17:34.408721 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rnftv" podStartSLOduration=21.998468914 podStartE2EDuration="28.408695871s" podCreationTimestamp="2026-03-20 16:17:06 +0000 UTC" firstStartedPulling="2026-03-20 16:17:26.411200197 +0000 UTC m=+997.357568012" lastFinishedPulling="2026-03-20 16:17:32.821427144 +0000 UTC m=+1003.767794969" observedRunningTime="2026-03-20 16:17:34.398164834 +0000 UTC m=+1005.344532649" watchObservedRunningTime="2026-03-20 16:17:34.408695871 +0000 UTC m=+1005.355063686" Mar 20 16:17:35 crc kubenswrapper[4936]: I0320 16:17:35.359903 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tp2gk" event={"ID":"a660d73e-dab8-4751-8219-c91f6017b574","Type":"ContainerStarted","Data":"4342a9d3c0c17fe337cfd181fbc126aad19f9e118fa0b7a9d6cd85e0acec7ddb"} Mar 20 16:17:35 crc kubenswrapper[4936]: I0320 16:17:35.360947 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:35 crc kubenswrapper[4936]: I0320 16:17:35.360967 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tp2gk" event={"ID":"a660d73e-dab8-4751-8219-c91f6017b574","Type":"ContainerStarted","Data":"bc8c6dc7ea89fc988018f9092d31a125d9481505eb862217f84f9da79bebefdb"} Mar 20 16:17:35 crc kubenswrapper[4936]: I0320 16:17:35.360988 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:17:35 crc kubenswrapper[4936]: I0320 16:17:35.362653 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" event={"ID":"e353d140-f7e5-47b4-b18e-200d5a255086","Type":"ContainerStarted","Data":"8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d"} Mar 20 16:17:35 crc kubenswrapper[4936]: I0320 16:17:35.362890 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-rnftv" Mar 20 16:17:35 crc kubenswrapper[4936]: I0320 16:17:35.393789 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-tp2gk" podStartSLOduration=24.305860096 podStartE2EDuration="29.393761678s" podCreationTimestamp="2026-03-20 16:17:06 +0000 UTC" firstStartedPulling="2026-03-20 16:17:26.552151955 +0000 UTC m=+997.498519770" lastFinishedPulling="2026-03-20 16:17:31.640053537 +0000 UTC m=+1002.586421352" observedRunningTime="2026-03-20 16:17:35.386729193 +0000 UTC m=+1006.333097008" watchObservedRunningTime="2026-03-20 16:17:35.393761678 +0000 UTC m=+1006.340129493" Mar 20 16:17:35 crc kubenswrapper[4936]: I0320 16:17:35.406651 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" podStartSLOduration=3.414443973 podStartE2EDuration="39.406627403s" podCreationTimestamp="2026-03-20 16:16:56 +0000 UTC" firstStartedPulling="2026-03-20 16:16:57.349570488 +0000 UTC m=+968.295938303" lastFinishedPulling="2026-03-20 16:17:33.341753918 +0000 UTC m=+1004.288121733" observedRunningTime="2026-03-20 16:17:35.401814143 +0000 UTC m=+1006.348181968" watchObservedRunningTime="2026-03-20 16:17:35.406627403 +0000 UTC m=+1006.352995228" Mar 20 16:17:37 crc kubenswrapper[4936]: I0320 16:17:37.041647 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:17:38 crc kubenswrapper[4936]: I0320 16:17:38.392745 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" event={"ID":"ce7feb82-b60a-4432-b1b9-b6a11f0e1839","Type":"ContainerStarted","Data":"00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905"} Mar 20 16:17:39 crc kubenswrapper[4936]: I0320 16:17:39.401836 4936 generic.go:334] "Generic (PLEG): container finished" podID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" containerID="00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905" exitCode=0 Mar 20 16:17:39 crc kubenswrapper[4936]: I0320 16:17:39.401903 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" event={"ID":"ce7feb82-b60a-4432-b1b9-b6a11f0e1839","Type":"ContainerDied","Data":"00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905"} Mar 20 16:17:39 crc kubenswrapper[4936]: I0320 16:17:39.597348 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 20 16:17:39 crc kubenswrapper[4936]: I0320 16:17:39.597428 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 20 16:17:41 crc kubenswrapper[4936]: I0320 16:17:41.285899 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:41 crc kubenswrapper[4936]: I0320 16:17:41.286157 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:41 crc kubenswrapper[4936]: I0320 16:17:41.571788 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 20 16:17:42 crc kubenswrapper[4936]: I0320 16:17:42.044826 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:17:42 crc kubenswrapper[4936]: I0320 16:17:42.151379 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mmbhh"] Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.079163 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.150185 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2d8d2157-625a-4b47-a722-1ccbe9236ef6" containerName="galera" probeResult="failure" output=< Mar 20 16:17:43 crc kubenswrapper[4936]: wsrep_local_state_comment (Joined) differs from Synced Mar 20 16:17:43 crc kubenswrapper[4936]: > Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.422199 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.431481 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-lsdbt"] Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.432806 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.443428 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-lsdbt"] Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.560062 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-config\") pod \"dnsmasq-dns-7cb5889db5-lsdbt\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.560132 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh4sm\" (UniqueName: \"kubernetes.io/projected/688fe90d-ac38-4e61-bbb0-936086990424-kube-api-access-fh4sm\") pod \"dnsmasq-dns-7cb5889db5-lsdbt\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.560190 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-lsdbt\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.661511 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh4sm\" (UniqueName: \"kubernetes.io/projected/688fe90d-ac38-4e61-bbb0-936086990424-kube-api-access-fh4sm\") pod \"dnsmasq-dns-7cb5889db5-lsdbt\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.661621 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-lsdbt\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.661740 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-config\") pod \"dnsmasq-dns-7cb5889db5-lsdbt\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.662496 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-lsdbt\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.662619 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-config\") pod \"dnsmasq-dns-7cb5889db5-lsdbt\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.690796 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh4sm\" (UniqueName: \"kubernetes.io/projected/688fe90d-ac38-4e61-bbb0-936086990424-kube-api-access-fh4sm\") pod \"dnsmasq-dns-7cb5889db5-lsdbt\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:43 crc kubenswrapper[4936]: I0320 16:17:43.757330 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.639576 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.645906 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.648508 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.649012 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-p96k5" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.649667 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.650492 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.663887 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.778135 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b1ea233b-41e7-49ac-a878-aa038b350252-cache\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.778223 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jmvf\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-kube-api-access-9jmvf\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.778314 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.778383 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ea233b-41e7-49ac-a878-aa038b350252-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.778425 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.778479 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b1ea233b-41e7-49ac-a878-aa038b350252-lock\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.880436 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b1ea233b-41e7-49ac-a878-aa038b350252-lock\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.880512 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b1ea233b-41e7-49ac-a878-aa038b350252-cache\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.880590 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jmvf\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-kube-api-access-9jmvf\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.880615 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.880659 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ea233b-41e7-49ac-a878-aa038b350252-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.880729 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: E0320 16:17:44.880867 4936 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 20 16:17:44 crc kubenswrapper[4936]: E0320 16:17:44.880881 4936 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 20 16:17:44 crc kubenswrapper[4936]: E0320 16:17:44.880924 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift podName:b1ea233b-41e7-49ac-a878-aa038b350252 nodeName:}" failed. No retries permitted until 2026-03-20 16:17:45.380907585 +0000 UTC m=+1016.327275400 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift") pod "swift-storage-0" (UID: "b1ea233b-41e7-49ac-a878-aa038b350252") : configmap "swift-ring-files" not found Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.881200 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.881290 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b1ea233b-41e7-49ac-a878-aa038b350252-cache\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.881302 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b1ea233b-41e7-49ac-a878-aa038b350252-lock\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.885386 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ea233b-41e7-49ac-a878-aa038b350252-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.898770 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jmvf\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-kube-api-access-9jmvf\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:44 crc kubenswrapper[4936]: I0320 16:17:44.903455 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:45 crc kubenswrapper[4936]: I0320 16:17:45.388559 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:45 crc kubenswrapper[4936]: E0320 16:17:45.388930 4936 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 20 16:17:45 crc kubenswrapper[4936]: E0320 16:17:45.388983 4936 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 20 16:17:45 crc kubenswrapper[4936]: E0320 16:17:45.389058 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift podName:b1ea233b-41e7-49ac-a878-aa038b350252 nodeName:}" failed. No retries permitted until 2026-03-20 16:17:46.389033413 +0000 UTC m=+1017.335401248 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift") pod "swift-storage-0" (UID: "b1ea233b-41e7-49ac-a878-aa038b350252") : configmap "swift-ring-files" not found Mar 20 16:17:46 crc kubenswrapper[4936]: I0320 16:17:46.404290 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:46 crc kubenswrapper[4936]: E0320 16:17:46.404532 4936 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 20 16:17:46 crc kubenswrapper[4936]: E0320 16:17:46.404602 4936 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 20 16:17:46 crc kubenswrapper[4936]: E0320 16:17:46.404687 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift podName:b1ea233b-41e7-49ac-a878-aa038b350252 nodeName:}" failed. No retries permitted until 2026-03-20 16:17:48.40465866 +0000 UTC m=+1019.351026495 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift") pod "swift-storage-0" (UID: "b1ea233b-41e7-49ac-a878-aa038b350252") : configmap "swift-ring-files" not found Mar 20 16:17:47 crc kubenswrapper[4936]: E0320 16:17:47.875683 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Mar 20 16:17:47 crc kubenswrapper[4936]: E0320 16:17:47.876252 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:n7bhf9hcch5ddh5c7h566h55bh5f8hbdh654h5d6h695h66bh676h65fh5b6h568h6fh569h5cchc4h5d4hbchd4h5c8h677h67hf5h5ffh68fh559h67q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9kkvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(7b0e8ad7-4143-431d-a022-c467ee2d09ac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:17:47 crc kubenswrapper[4936]: E0320 16:17:47.877319 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="7b0e8ad7-4143-431d-a022-c467ee2d09ac" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.242983 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.325162 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.437950 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:48 crc kubenswrapper[4936]: E0320 16:17:48.438069 4936 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 20 16:17:48 crc kubenswrapper[4936]: E0320 16:17:48.438085 4936 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 20 16:17:48 crc kubenswrapper[4936]: E0320 16:17:48.438138 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift podName:b1ea233b-41e7-49ac-a878-aa038b350252 nodeName:}" failed. No retries permitted until 2026-03-20 16:17:52.438112989 +0000 UTC m=+1023.384480804 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift") pod "swift-storage-0" (UID: "b1ea233b-41e7-49ac-a878-aa038b350252") : configmap "swift-ring-files" not found Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.471157 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7","Type":"ContainerStarted","Data":"cc65f008cae29d2a1455317a25744460d8b8e012a37d4ecac0d08db64f7e8ca1"} Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.475967 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" podUID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" containerName="dnsmasq-dns" containerID="cri-o://e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77" gracePeriod=10 Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.476318 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" event={"ID":"ce7feb82-b60a-4432-b1b9-b6a11f0e1839","Type":"ContainerStarted","Data":"e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77"} Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.476358 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.497773 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=18.031312249 podStartE2EDuration="39.497747747s" podCreationTimestamp="2026-03-20 16:17:09 +0000 UTC" firstStartedPulling="2026-03-20 16:17:26.647493013 +0000 UTC m=+997.593860828" lastFinishedPulling="2026-03-20 16:17:48.113928511 +0000 UTC m=+1019.060296326" observedRunningTime="2026-03-20 16:17:48.489816236 +0000 UTC m=+1019.436184061" watchObservedRunningTime="2026-03-20 16:17:48.497747747 +0000 UTC m=+1019.444115562" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.537128 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-lsdbt"] Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.538719 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" podStartSLOduration=-9223371984.316072 podStartE2EDuration="52.5387026s" podCreationTimestamp="2026-03-20 16:16:56 +0000 UTC" firstStartedPulling="2026-03-20 16:16:57.133236325 +0000 UTC m=+968.079604140" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:17:48.527349419 +0000 UTC m=+1019.473717234" watchObservedRunningTime="2026-03-20 16:17:48.5387026 +0000 UTC m=+1019.485070415" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.597304 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6k8ph"] Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.598685 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.603513 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.607627 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.607823 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.612391 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6k8ph"] Mar 20 16:17:48 crc kubenswrapper[4936]: W0320 16:17:48.624705 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod688fe90d_ac38_4e61_bbb0_936086990424.slice/crio-ce26a09c1caaa69fb7ef0e786ac632a9fe5815d60a6ad10c2eefff9d60dc9d79 WatchSource:0}: Error finding container ce26a09c1caaa69fb7ef0e786ac632a9fe5815d60a6ad10c2eefff9d60dc9d79: Status 404 returned error can't find the container with id ce26a09c1caaa69fb7ef0e786ac632a9fe5815d60a6ad10c2eefff9d60dc9d79 Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.643446 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-scripts\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.643553 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/070d6439-9bd3-4fd8-b61a-df1f03badb79-etc-swift\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.643572 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-ring-data-devices\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.643604 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-dispersionconf\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.643623 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-combined-ca-bundle\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.643715 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdvw9\" (UniqueName: \"kubernetes.io/projected/070d6439-9bd3-4fd8-b61a-df1f03badb79-kube-api-access-wdvw9\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.643864 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-swiftconf\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.750231 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-scripts\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.750825 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/070d6439-9bd3-4fd8-b61a-df1f03badb79-etc-swift\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.750850 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-ring-data-devices\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.750889 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-dispersionconf\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.750920 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-combined-ca-bundle\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.750954 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdvw9\" (UniqueName: \"kubernetes.io/projected/070d6439-9bd3-4fd8-b61a-df1f03badb79-kube-api-access-wdvw9\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.751041 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-swiftconf\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.751322 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-scripts\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.751840 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/070d6439-9bd3-4fd8-b61a-df1f03badb79-etc-swift\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.752588 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-ring-data-devices\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.759058 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-combined-ca-bundle\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.759770 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-dispersionconf\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.762433 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-swiftconf\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.778164 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdvw9\" (UniqueName: \"kubernetes.io/projected/070d6439-9bd3-4fd8-b61a-df1f03badb79-kube-api-access-wdvw9\") pod \"swift-ring-rebalance-6k8ph\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.869233 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.950717 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.985736 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-config\") pod \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.985782 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-dns-svc\") pod \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " Mar 20 16:17:48 crc kubenswrapper[4936]: I0320 16:17:48.985827 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fm6x\" (UniqueName: \"kubernetes.io/projected/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-kube-api-access-2fm6x\") pod \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\" (UID: \"ce7feb82-b60a-4432-b1b9-b6a11f0e1839\") " Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.020825 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-kube-api-access-2fm6x" (OuterVolumeSpecName: "kube-api-access-2fm6x") pod "ce7feb82-b60a-4432-b1b9-b6a11f0e1839" (UID: "ce7feb82-b60a-4432-b1b9-b6a11f0e1839"). InnerVolumeSpecName "kube-api-access-2fm6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.063966 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce7feb82-b60a-4432-b1b9-b6a11f0e1839" (UID: "ce7feb82-b60a-4432-b1b9-b6a11f0e1839"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.063980 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-config" (OuterVolumeSpecName: "config") pod "ce7feb82-b60a-4432-b1b9-b6a11f0e1839" (UID: "ce7feb82-b60a-4432-b1b9-b6a11f0e1839"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.090852 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fm6x\" (UniqueName: \"kubernetes.io/projected/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-kube-api-access-2fm6x\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.090892 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.090903 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce7feb82-b60a-4432-b1b9-b6a11f0e1839-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.438036 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.460743 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6k8ph"] Mar 20 16:17:49 crc kubenswrapper[4936]: W0320 16:17:49.466832 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod070d6439_9bd3_4fd8_b61a_df1f03badb79.slice/crio-85d0a6cc208da0c19ae0a54a88c4f6c2e0a9f0ec7b253595f38eef9840ae6ae9 WatchSource:0}: Error finding container 85d0a6cc208da0c19ae0a54a88c4f6c2e0a9f0ec7b253595f38eef9840ae6ae9: Status 404 returned error can't find the container with id 85d0a6cc208da0c19ae0a54a88c4f6c2e0a9f0ec7b253595f38eef9840ae6ae9 Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.494394 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.495418 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7b0e8ad7-4143-431d-a022-c467ee2d09ac","Type":"ContainerStarted","Data":"7715263cafc967954e6d0f735d14c467cebea127f5e1dc5cda59ebdfbbf22fe8"} Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.499364 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6k8ph" event={"ID":"070d6439-9bd3-4fd8-b61a-df1f03badb79","Type":"ContainerStarted","Data":"85d0a6cc208da0c19ae0a54a88c4f6c2e0a9f0ec7b253595f38eef9840ae6ae9"} Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.501175 4936 generic.go:334] "Generic (PLEG): container finished" podID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" containerID="e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77" exitCode=0 Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.501221 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" event={"ID":"ce7feb82-b60a-4432-b1b9-b6a11f0e1839","Type":"ContainerDied","Data":"e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77"} Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.501240 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" event={"ID":"ce7feb82-b60a-4432-b1b9-b6a11f0e1839","Type":"ContainerDied","Data":"d8318d40d0f80f5decd1ad62ac6313c77c32f02041da5d409897cdd2ceaba73b"} Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.501258 4936 scope.go:117] "RemoveContainer" containerID="e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.501367 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mmbhh" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.507582 4936 generic.go:334] "Generic (PLEG): container finished" podID="688fe90d-ac38-4e61-bbb0-936086990424" containerID="a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1" exitCode=0 Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.507707 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" event={"ID":"688fe90d-ac38-4e61-bbb0-936086990424","Type":"ContainerDied","Data":"a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1"} Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.507817 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" event={"ID":"688fe90d-ac38-4e61-bbb0-936086990424","Type":"ContainerStarted","Data":"ce26a09c1caaa69fb7ef0e786ac632a9fe5815d60a6ad10c2eefff9d60dc9d79"} Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.508429 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.559319 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.566161 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=37.942150928 podStartE2EDuration="43.566126692s" podCreationTimestamp="2026-03-20 16:17:06 +0000 UTC" firstStartedPulling="2026-03-20 16:17:27.188400396 +0000 UTC m=+998.134768211" lastFinishedPulling="2026-03-20 16:17:32.81237617 +0000 UTC m=+1003.758743975" observedRunningTime="2026-03-20 16:17:49.558647624 +0000 UTC m=+1020.505015449" watchObservedRunningTime="2026-03-20 16:17:49.566126692 +0000 UTC m=+1020.512494527" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.735041 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-zv6gp"] Mar 20 16:17:49 crc kubenswrapper[4936]: E0320 16:17:49.735711 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" containerName="init" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.735820 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" containerName="init" Mar 20 16:17:49 crc kubenswrapper[4936]: E0320 16:17:49.735929 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" containerName="dnsmasq-dns" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.736006 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" containerName="dnsmasq-dns" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.736280 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" containerName="dnsmasq-dns" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.737033 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.741513 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.754257 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-zv6gp"] Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.804605 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd16e0f6-1913-4750-adad-8db14435dd01-operator-scripts\") pod \"root-account-create-update-zv6gp\" (UID: \"cd16e0f6-1913-4750-adad-8db14435dd01\") " pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.804838 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcp5t\" (UniqueName: \"kubernetes.io/projected/cd16e0f6-1913-4750-adad-8db14435dd01-kube-api-access-dcp5t\") pod \"root-account-create-update-zv6gp\" (UID: \"cd16e0f6-1913-4750-adad-8db14435dd01\") " pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.866985 4936 scope.go:117] "RemoveContainer" containerID="00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.915173 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd16e0f6-1913-4750-adad-8db14435dd01-operator-scripts\") pod \"root-account-create-update-zv6gp\" (UID: \"cd16e0f6-1913-4750-adad-8db14435dd01\") " pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.915227 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcp5t\" (UniqueName: \"kubernetes.io/projected/cd16e0f6-1913-4750-adad-8db14435dd01-kube-api-access-dcp5t\") pod \"root-account-create-update-zv6gp\" (UID: \"cd16e0f6-1913-4750-adad-8db14435dd01\") " pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.916161 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd16e0f6-1913-4750-adad-8db14435dd01-operator-scripts\") pod \"root-account-create-update-zv6gp\" (UID: \"cd16e0f6-1913-4750-adad-8db14435dd01\") " pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.920359 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.920391 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-lsdbt"] Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.920407 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-ggvg7"] Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.926174 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-ggvg7"] Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.934946 4936 scope.go:117] "RemoveContainer" containerID="e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77" Mar 20 16:17:49 crc kubenswrapper[4936]: E0320 16:17:49.935329 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77\": container with ID starting with e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77 not found: ID does not exist" containerID="e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.935360 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77"} err="failed to get container status \"e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77\": rpc error: code = NotFound desc = could not find container \"e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77\": container with ID starting with e572a05714a65dc629bee0321ed2787d1477292b4d743cfd67f347ae086def77 not found: ID does not exist" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.935381 4936 scope.go:117] "RemoveContainer" containerID="00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905" Mar 20 16:17:49 crc kubenswrapper[4936]: E0320 16:17:49.935649 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905\": container with ID starting with 00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905 not found: ID does not exist" containerID="00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.935858 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905"} err="failed to get container status \"00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905\": rpc error: code = NotFound desc = could not find container \"00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905\": container with ID starting with 00665d8b8a70fa70cdfbc9cc7defbaf1211dbed1b53105f3a2b856031aa12905 not found: ID does not exist" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.939720 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcp5t\" (UniqueName: \"kubernetes.io/projected/cd16e0f6-1913-4750-adad-8db14435dd01-kube-api-access-dcp5t\") pod \"root-account-create-update-zv6gp\" (UID: \"cd16e0f6-1913-4750-adad-8db14435dd01\") " pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.944567 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.948447 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 20 16:17:49 crc kubenswrapper[4936]: I0320 16:17:49.990218 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mmbhh"] Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.016471 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-config\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.016825 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.016992 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znxsc\" (UniqueName: \"kubernetes.io/projected/f377f06a-3265-4b2a-8a33-1e39a572ecb0-kube-api-access-znxsc\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.017052 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.031472 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mmbhh"] Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.041691 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-vpzks"] Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.043165 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.045926 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.052712 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.079846 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vpzks"] Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.098449 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.120213 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhf6d\" (UniqueName: \"kubernetes.io/projected/d6b58590-f075-4e5c-8bad-6f8ebb368c93-kube-api-access-lhf6d\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.120268 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d6b58590-f075-4e5c-8bad-6f8ebb368c93-ovs-rundir\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.120315 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6b58590-f075-4e5c-8bad-6f8ebb368c93-config\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.120522 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-config\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.120681 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d6b58590-f075-4e5c-8bad-6f8ebb368c93-ovn-rundir\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.120764 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b58590-f075-4e5c-8bad-6f8ebb368c93-combined-ca-bundle\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.121058 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.121126 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6b58590-f075-4e5c-8bad-6f8ebb368c93-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.121238 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znxsc\" (UniqueName: \"kubernetes.io/projected/f377f06a-3265-4b2a-8a33-1e39a572ecb0-kube-api-access-znxsc\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.121305 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.122179 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.132822 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-config\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.133309 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.141316 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znxsc\" (UniqueName: \"kubernetes.io/projected/f377f06a-3265-4b2a-8a33-1e39a572ecb0-kube-api-access-znxsc\") pod \"dnsmasq-dns-74f6f696b9-ggvg7\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.214249 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.222855 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6b58590-f075-4e5c-8bad-6f8ebb368c93-config\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.222948 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d6b58590-f075-4e5c-8bad-6f8ebb368c93-ovn-rundir\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.223008 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b58590-f075-4e5c-8bad-6f8ebb368c93-combined-ca-bundle\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.223123 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6b58590-f075-4e5c-8bad-6f8ebb368c93-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.223247 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhf6d\" (UniqueName: \"kubernetes.io/projected/d6b58590-f075-4e5c-8bad-6f8ebb368c93-kube-api-access-lhf6d\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.223275 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d6b58590-f075-4e5c-8bad-6f8ebb368c93-ovs-rundir\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.223443 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d6b58590-f075-4e5c-8bad-6f8ebb368c93-ovs-rundir\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.223752 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d6b58590-f075-4e5c-8bad-6f8ebb368c93-ovn-rundir\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.224197 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6b58590-f075-4e5c-8bad-6f8ebb368c93-config\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.228893 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-ggvg7"] Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.229607 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.238899 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6b58590-f075-4e5c-8bad-6f8ebb368c93-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.257876 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b58590-f075-4e5c-8bad-6f8ebb368c93-combined-ca-bundle\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.264132 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhf6d\" (UniqueName: \"kubernetes.io/projected/d6b58590-f075-4e5c-8bad-6f8ebb368c93-kube-api-access-lhf6d\") pod \"ovn-controller-metrics-vpzks\" (UID: \"d6b58590-f075-4e5c-8bad-6f8ebb368c93\") " pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.264815 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-22blq"] Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.284960 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-22blq"] Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.285079 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.289438 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.325437 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.325503 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.325578 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-config\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.325598 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzdsk\" (UniqueName: \"kubernetes.io/projected/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-kube-api-access-wzdsk\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.325680 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-dns-svc\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.368504 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vpzks" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.426709 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.426798 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-config\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.426824 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzdsk\" (UniqueName: \"kubernetes.io/projected/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-kube-api-access-wzdsk\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.426898 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-dns-svc\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.426950 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.427741 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.428362 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.429069 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-config\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.430720 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-dns-svc\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.448151 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzdsk\" (UniqueName: \"kubernetes.io/projected/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-kube-api-access-wzdsk\") pod \"dnsmasq-dns-698758b865-22blq\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.552988 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18cfd431-66b8-49d8-8dd9-b2184415c006","Type":"ContainerStarted","Data":"3420f7533e5b46917dbbbf04a1c49349ccdcf30014af6ad50b99944a468f9184"} Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.561935 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b3a9af8-80db-4edc-9f6a-42b641c25d71","Type":"ContainerStarted","Data":"5f14a26871add7074b92241d4cf3fd824a81b961fa684c3affe2d0ef82a2d850"} Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.569003 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" event={"ID":"688fe90d-ac38-4e61-bbb0-936086990424","Type":"ContainerStarted","Data":"0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca"} Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.569441 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.569428 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" podUID="688fe90d-ac38-4e61-bbb0-936086990424" containerName="dnsmasq-dns" containerID="cri-o://0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca" gracePeriod=10 Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.633108 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" podStartSLOduration=7.633058374 podStartE2EDuration="7.633058374s" podCreationTimestamp="2026-03-20 16:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:17:50.627599284 +0000 UTC m=+1021.573967119" watchObservedRunningTime="2026-03-20 16:17:50.633058374 +0000 UTC m=+1021.579426189" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.644434 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.686011 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.823567 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-zv6gp"] Mar 20 16:17:50 crc kubenswrapper[4936]: W0320 16:17:50.848786 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd16e0f6_1913_4750_adad_8db14435dd01.slice/crio-4572eb60641e39587be859e399f036ed489cab2de2d4457653df779ef2febf51 WatchSource:0}: Error finding container 4572eb60641e39587be859e399f036ed489cab2de2d4457653df779ef2febf51: Status 404 returned error can't find the container with id 4572eb60641e39587be859e399f036ed489cab2de2d4457653df779ef2febf51 Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.861233 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.940223 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.941723 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 20 16:17:50 crc kubenswrapper[4936]: I0320 16:17:50.964962 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-z78gp" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.001208 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.001461 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.003968 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.018030 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.048037 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-ggvg7"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.087501 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vpzks"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.164630 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f4f09dd-f403-4f3b-ac44-129df52fd796-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.164760 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7f4f09dd-f403-4f3b-ac44-129df52fd796-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.164796 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4f09dd-f403-4f3b-ac44-129df52fd796-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.164821 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f4f09dd-f403-4f3b-ac44-129df52fd796-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.164990 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f4f09dd-f403-4f3b-ac44-129df52fd796-scripts\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.165023 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4f09dd-f403-4f3b-ac44-129df52fd796-config\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.165041 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx5n6\" (UniqueName: \"kubernetes.io/projected/7f4f09dd-f403-4f3b-ac44-129df52fd796-kube-api-access-qx5n6\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.266012 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f4f09dd-f403-4f3b-ac44-129df52fd796-scripts\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.266310 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4f09dd-f403-4f3b-ac44-129df52fd796-config\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.266327 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx5n6\" (UniqueName: \"kubernetes.io/projected/7f4f09dd-f403-4f3b-ac44-129df52fd796-kube-api-access-qx5n6\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.266383 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f4f09dd-f403-4f3b-ac44-129df52fd796-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.266429 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7f4f09dd-f403-4f3b-ac44-129df52fd796-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.266448 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4f09dd-f403-4f3b-ac44-129df52fd796-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.266470 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f4f09dd-f403-4f3b-ac44-129df52fd796-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.268729 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7f4f09dd-f403-4f3b-ac44-129df52fd796-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.268729 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f4f09dd-f403-4f3b-ac44-129df52fd796-scripts\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.268925 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4f09dd-f403-4f3b-ac44-129df52fd796-config\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.271743 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f4f09dd-f403-4f3b-ac44-129df52fd796-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.274036 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4f09dd-f403-4f3b-ac44-129df52fd796-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.275260 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f4f09dd-f403-4f3b-ac44-129df52fd796-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.287220 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx5n6\" (UniqueName: \"kubernetes.io/projected/7f4f09dd-f403-4f3b-ac44-129df52fd796-kube-api-access-qx5n6\") pod \"ovn-northd-0\" (UID: \"7f4f09dd-f403-4f3b-ac44-129df52fd796\") " pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.306982 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-22blq"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.322752 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.365169 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.367246 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-config\") pod \"688fe90d-ac38-4e61-bbb0-936086990424\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.367283 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh4sm\" (UniqueName: \"kubernetes.io/projected/688fe90d-ac38-4e61-bbb0-936086990424-kube-api-access-fh4sm\") pod \"688fe90d-ac38-4e61-bbb0-936086990424\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.367392 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-dns-svc\") pod \"688fe90d-ac38-4e61-bbb0-936086990424\" (UID: \"688fe90d-ac38-4e61-bbb0-936086990424\") " Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.372692 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/688fe90d-ac38-4e61-bbb0-936086990424-kube-api-access-fh4sm" (OuterVolumeSpecName: "kube-api-access-fh4sm") pod "688fe90d-ac38-4e61-bbb0-936086990424" (UID: "688fe90d-ac38-4e61-bbb0-936086990424"). InnerVolumeSpecName "kube-api-access-fh4sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.472088 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh4sm\" (UniqueName: \"kubernetes.io/projected/688fe90d-ac38-4e61-bbb0-936086990424-kube-api-access-fh4sm\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.486187 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-config" (OuterVolumeSpecName: "config") pod "688fe90d-ac38-4e61-bbb0-936086990424" (UID: "688fe90d-ac38-4e61-bbb0-936086990424"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.494997 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "688fe90d-ac38-4e61-bbb0-936086990424" (UID: "688fe90d-ac38-4e61-bbb0-936086990424"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.573559 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.573590 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/688fe90d-ac38-4e61-bbb0-936086990424-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.580369 4936 generic.go:334] "Generic (PLEG): container finished" podID="cd16e0f6-1913-4750-adad-8db14435dd01" containerID="cf0c7f5c9b08378d74801d46ad16cbd02d8d77f2d374525945836e7538e47ffd" exitCode=0 Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.580453 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zv6gp" event={"ID":"cd16e0f6-1913-4750-adad-8db14435dd01","Type":"ContainerDied","Data":"cf0c7f5c9b08378d74801d46ad16cbd02d8d77f2d374525945836e7538e47ffd"} Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.580486 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zv6gp" event={"ID":"cd16e0f6-1913-4750-adad-8db14435dd01","Type":"ContainerStarted","Data":"4572eb60641e39587be859e399f036ed489cab2de2d4457653df779ef2febf51"} Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.587835 4936 generic.go:334] "Generic (PLEG): container finished" podID="688fe90d-ac38-4e61-bbb0-936086990424" containerID="0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca" exitCode=0 Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.587893 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" event={"ID":"688fe90d-ac38-4e61-bbb0-936086990424","Type":"ContainerDied","Data":"0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca"} Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.587910 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" event={"ID":"688fe90d-ac38-4e61-bbb0-936086990424","Type":"ContainerDied","Data":"ce26a09c1caaa69fb7ef0e786ac632a9fe5815d60a6ad10c2eefff9d60dc9d79"} Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.587926 4936 scope.go:117] "RemoveContainer" containerID="0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.588034 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-lsdbt" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.590390 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-22blq" event={"ID":"223bcfee-4f1a-4ce5-9e38-a8bd4676411b","Type":"ContainerStarted","Data":"162fc1cbc8c40839a51a85bd28799da36f43202906d9030242282e0ca56a8d9f"} Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.612816 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vpzks" event={"ID":"d6b58590-f075-4e5c-8bad-6f8ebb368c93","Type":"ContainerStarted","Data":"37bd054ab026397734e31ad3c8b984595318c83f6402dc89335dc3813aa18c48"} Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.612884 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vpzks" event={"ID":"d6b58590-f075-4e5c-8bad-6f8ebb368c93","Type":"ContainerStarted","Data":"7617103f0dc38274f4b2c597481314d265580ced9ecf963eb53e9820ca43368c"} Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.616837 4936 generic.go:334] "Generic (PLEG): container finished" podID="f377f06a-3265-4b2a-8a33-1e39a572ecb0" containerID="4937fd5d5ca10717ae68151bac3416a90dc4f6e937df093f6a4e2ee952bd6ece" exitCode=0 Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.617183 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" event={"ID":"f377f06a-3265-4b2a-8a33-1e39a572ecb0","Type":"ContainerDied","Data":"4937fd5d5ca10717ae68151bac3416a90dc4f6e937df093f6a4e2ee952bd6ece"} Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.617221 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" event={"ID":"f377f06a-3265-4b2a-8a33-1e39a572ecb0","Type":"ContainerStarted","Data":"09579b62ab8fe01c9affe9f42584b7645487b92d8cf87c50174ed6c1f3b91e12"} Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.641484 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-vpzks" podStartSLOduration=2.641460571 podStartE2EDuration="2.641460571s" podCreationTimestamp="2026-03-20 16:17:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:17:51.634281101 +0000 UTC m=+1022.580648926" watchObservedRunningTime="2026-03-20 16:17:51.641460571 +0000 UTC m=+1022.587828386" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.647334 4936 scope.go:117] "RemoveContainer" containerID="a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.696934 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-lsdbt"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.708239 4936 scope.go:117] "RemoveContainer" containerID="0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca" Mar 20 16:17:51 crc kubenswrapper[4936]: E0320 16:17:51.709771 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca\": container with ID starting with 0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca not found: ID does not exist" containerID="0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.709817 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca"} err="failed to get container status \"0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca\": rpc error: code = NotFound desc = could not find container \"0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca\": container with ID starting with 0f7eb543cbb129c8a5050c9c3567b0ceaf16e9c419c3741cae9d1457dbdd43ca not found: ID does not exist" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.709845 4936 scope.go:117] "RemoveContainer" containerID="a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1" Mar 20 16:17:51 crc kubenswrapper[4936]: E0320 16:17:51.710734 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1\": container with ID starting with a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1 not found: ID does not exist" containerID="a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.710834 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1"} err="failed to get container status \"a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1\": rpc error: code = NotFound desc = could not find container \"a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1\": container with ID starting with a133c75ae1d0266064c87a3be558af83d340af41937056f29cbbf63db2693bb1 not found: ID does not exist" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.711387 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-lsdbt"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.824758 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.838581 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-9vpbv"] Mar 20 16:17:51 crc kubenswrapper[4936]: E0320 16:17:51.838930 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688fe90d-ac38-4e61-bbb0-936086990424" containerName="dnsmasq-dns" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.838944 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="688fe90d-ac38-4e61-bbb0-936086990424" containerName="dnsmasq-dns" Mar 20 16:17:51 crc kubenswrapper[4936]: E0320 16:17:51.838993 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688fe90d-ac38-4e61-bbb0-936086990424" containerName="init" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.839000 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="688fe90d-ac38-4e61-bbb0-936086990424" containerName="init" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.839177 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="688fe90d-ac38-4e61-bbb0-936086990424" containerName="dnsmasq-dns" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.839759 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.850238 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9vpbv"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.872575 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="688fe90d-ac38-4e61-bbb0-936086990424" path="/var/lib/kubelet/pods/688fe90d-ac38-4e61-bbb0-936086990424/volumes" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.873213 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce7feb82-b60a-4432-b1b9-b6a11f0e1839" path="/var/lib/kubelet/pods/ce7feb82-b60a-4432-b1b9-b6a11f0e1839/volumes" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.873798 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-05f5-account-create-update-9vd8r"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.874900 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.877559 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.886170 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-05f5-account-create-update-9vd8r"] Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.981671 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/812b9782-84b1-41fc-939e-30e82c1b5e13-operator-scripts\") pod \"glance-db-create-9vpbv\" (UID: \"812b9782-84b1-41fc-939e-30e82c1b5e13\") " pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.981748 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-operator-scripts\") pod \"glance-05f5-account-create-update-9vd8r\" (UID: \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\") " pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.981790 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2v8h\" (UniqueName: \"kubernetes.io/projected/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-kube-api-access-n2v8h\") pod \"glance-05f5-account-create-update-9vd8r\" (UID: \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\") " pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:51 crc kubenswrapper[4936]: I0320 16:17:51.981868 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxxmw\" (UniqueName: \"kubernetes.io/projected/812b9782-84b1-41fc-939e-30e82c1b5e13-kube-api-access-jxxmw\") pod \"glance-db-create-9vpbv\" (UID: \"812b9782-84b1-41fc-939e-30e82c1b5e13\") " pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.041754 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-zphvf"] Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.043517 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.051112 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zphvf"] Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.083469 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxxmw\" (UniqueName: \"kubernetes.io/projected/812b9782-84b1-41fc-939e-30e82c1b5e13-kube-api-access-jxxmw\") pod \"glance-db-create-9vpbv\" (UID: \"812b9782-84b1-41fc-939e-30e82c1b5e13\") " pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.083651 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/812b9782-84b1-41fc-939e-30e82c1b5e13-operator-scripts\") pod \"glance-db-create-9vpbv\" (UID: \"812b9782-84b1-41fc-939e-30e82c1b5e13\") " pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.083675 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-operator-scripts\") pod \"glance-05f5-account-create-update-9vd8r\" (UID: \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\") " pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.083697 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2v8h\" (UniqueName: \"kubernetes.io/projected/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-kube-api-access-n2v8h\") pod \"glance-05f5-account-create-update-9vd8r\" (UID: \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\") " pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.084316 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/812b9782-84b1-41fc-939e-30e82c1b5e13-operator-scripts\") pod \"glance-db-create-9vpbv\" (UID: \"812b9782-84b1-41fc-939e-30e82c1b5e13\") " pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.084581 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-operator-scripts\") pod \"glance-05f5-account-create-update-9vd8r\" (UID: \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\") " pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.103680 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxxmw\" (UniqueName: \"kubernetes.io/projected/812b9782-84b1-41fc-939e-30e82c1b5e13-kube-api-access-jxxmw\") pod \"glance-db-create-9vpbv\" (UID: \"812b9782-84b1-41fc-939e-30e82c1b5e13\") " pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.111454 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2v8h\" (UniqueName: \"kubernetes.io/projected/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-kube-api-access-n2v8h\") pod \"glance-05f5-account-create-update-9vd8r\" (UID: \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\") " pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.145573 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7948-account-create-update-8trc2"] Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.147784 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.149787 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.155126 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7948-account-create-update-8trc2"] Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.171318 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.184695 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/434fff7b-3810-45b8-8028-98a9d77d3f15-operator-scripts\") pod \"keystone-db-create-zphvf\" (UID: \"434fff7b-3810-45b8-8028-98a9d77d3f15\") " pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.184822 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xccl\" (UniqueName: \"kubernetes.io/projected/434fff7b-3810-45b8-8028-98a9d77d3f15-kube-api-access-8xccl\") pod \"keystone-db-create-zphvf\" (UID: \"434fff7b-3810-45b8-8028-98a9d77d3f15\") " pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.198319 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.246195 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-jvq45"] Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.247514 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jvq45" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.250049 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jvq45"] Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.290136 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/434fff7b-3810-45b8-8028-98a9d77d3f15-operator-scripts\") pod \"keystone-db-create-zphvf\" (UID: \"434fff7b-3810-45b8-8028-98a9d77d3f15\") " pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.290338 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xccl\" (UniqueName: \"kubernetes.io/projected/434fff7b-3810-45b8-8028-98a9d77d3f15-kube-api-access-8xccl\") pod \"keystone-db-create-zphvf\" (UID: \"434fff7b-3810-45b8-8028-98a9d77d3f15\") " pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.290444 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0aa18c2-d52b-4671-996b-985c848a748f-operator-scripts\") pod \"keystone-7948-account-create-update-8trc2\" (UID: \"a0aa18c2-d52b-4671-996b-985c848a748f\") " pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.290637 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr4jb\" (UniqueName: \"kubernetes.io/projected/a0aa18c2-d52b-4671-996b-985c848a748f-kube-api-access-rr4jb\") pod \"keystone-7948-account-create-update-8trc2\" (UID: \"a0aa18c2-d52b-4671-996b-985c848a748f\") " pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.291481 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/434fff7b-3810-45b8-8028-98a9d77d3f15-operator-scripts\") pod \"keystone-db-create-zphvf\" (UID: \"434fff7b-3810-45b8-8028-98a9d77d3f15\") " pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.311824 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xccl\" (UniqueName: \"kubernetes.io/projected/434fff7b-3810-45b8-8028-98a9d77d3f15-kube-api-access-8xccl\") pod \"keystone-db-create-zphvf\" (UID: \"434fff7b-3810-45b8-8028-98a9d77d3f15\") " pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.364364 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.366516 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-962c-account-create-update-nbrwz"] Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.368035 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.370469 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.390385 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-962c-account-create-update-nbrwz"] Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.392213 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr4jb\" (UniqueName: \"kubernetes.io/projected/a0aa18c2-d52b-4671-996b-985c848a748f-kube-api-access-rr4jb\") pod \"keystone-7948-account-create-update-8trc2\" (UID: \"a0aa18c2-d52b-4671-996b-985c848a748f\") " pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.392351 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0aa18c2-d52b-4671-996b-985c848a748f-operator-scripts\") pod \"keystone-7948-account-create-update-8trc2\" (UID: \"a0aa18c2-d52b-4671-996b-985c848a748f\") " pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.392399 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8df2114-9d0b-43c8-91cd-a60355f7238b-operator-scripts\") pod \"placement-db-create-jvq45\" (UID: \"b8df2114-9d0b-43c8-91cd-a60355f7238b\") " pod="openstack/placement-db-create-jvq45" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.392425 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqrh9\" (UniqueName: \"kubernetes.io/projected/b8df2114-9d0b-43c8-91cd-a60355f7238b-kube-api-access-xqrh9\") pod \"placement-db-create-jvq45\" (UID: \"b8df2114-9d0b-43c8-91cd-a60355f7238b\") " pod="openstack/placement-db-create-jvq45" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.393436 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0aa18c2-d52b-4671-996b-985c848a748f-operator-scripts\") pod \"keystone-7948-account-create-update-8trc2\" (UID: \"a0aa18c2-d52b-4671-996b-985c848a748f\") " pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.428781 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr4jb\" (UniqueName: \"kubernetes.io/projected/a0aa18c2-d52b-4671-996b-985c848a748f-kube-api-access-rr4jb\") pod \"keystone-7948-account-create-update-8trc2\" (UID: \"a0aa18c2-d52b-4671-996b-985c848a748f\") " pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.494350 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8df2114-9d0b-43c8-91cd-a60355f7238b-operator-scripts\") pod \"placement-db-create-jvq45\" (UID: \"b8df2114-9d0b-43c8-91cd-a60355f7238b\") " pod="openstack/placement-db-create-jvq45" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.494696 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqrh9\" (UniqueName: \"kubernetes.io/projected/b8df2114-9d0b-43c8-91cd-a60355f7238b-kube-api-access-xqrh9\") pod \"placement-db-create-jvq45\" (UID: \"b8df2114-9d0b-43c8-91cd-a60355f7238b\") " pod="openstack/placement-db-create-jvq45" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.495007 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st8d5\" (UniqueName: \"kubernetes.io/projected/c48f3728-76d9-43c5-8053-b27ac3751ecb-kube-api-access-st8d5\") pod \"placement-962c-account-create-update-nbrwz\" (UID: \"c48f3728-76d9-43c5-8053-b27ac3751ecb\") " pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.495145 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48f3728-76d9-43c5-8053-b27ac3751ecb-operator-scripts\") pod \"placement-962c-account-create-update-nbrwz\" (UID: \"c48f3728-76d9-43c5-8053-b27ac3751ecb\") " pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.495271 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:17:52 crc kubenswrapper[4936]: E0320 16:17:52.496148 4936 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 20 16:17:52 crc kubenswrapper[4936]: E0320 16:17:52.496193 4936 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 20 16:17:52 crc kubenswrapper[4936]: E0320 16:17:52.498011 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift podName:b1ea233b-41e7-49ac-a878-aa038b350252 nodeName:}" failed. No retries permitted until 2026-03-20 16:18:00.497981232 +0000 UTC m=+1031.444349047 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift") pod "swift-storage-0" (UID: "b1ea233b-41e7-49ac-a878-aa038b350252") : configmap "swift-ring-files" not found Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.499465 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8df2114-9d0b-43c8-91cd-a60355f7238b-operator-scripts\") pod \"placement-db-create-jvq45\" (UID: \"b8df2114-9d0b-43c8-91cd-a60355f7238b\") " pod="openstack/placement-db-create-jvq45" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.515801 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqrh9\" (UniqueName: \"kubernetes.io/projected/b8df2114-9d0b-43c8-91cd-a60355f7238b-kube-api-access-xqrh9\") pod \"placement-db-create-jvq45\" (UID: \"b8df2114-9d0b-43c8-91cd-a60355f7238b\") " pod="openstack/placement-db-create-jvq45" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.534663 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.571109 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jvq45" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.600117 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st8d5\" (UniqueName: \"kubernetes.io/projected/c48f3728-76d9-43c5-8053-b27ac3751ecb-kube-api-access-st8d5\") pod \"placement-962c-account-create-update-nbrwz\" (UID: \"c48f3728-76d9-43c5-8053-b27ac3751ecb\") " pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.600173 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48f3728-76d9-43c5-8053-b27ac3751ecb-operator-scripts\") pod \"placement-962c-account-create-update-nbrwz\" (UID: \"c48f3728-76d9-43c5-8053-b27ac3751ecb\") " pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.601041 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48f3728-76d9-43c5-8053-b27ac3751ecb-operator-scripts\") pod \"placement-962c-account-create-update-nbrwz\" (UID: \"c48f3728-76d9-43c5-8053-b27ac3751ecb\") " pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.618561 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st8d5\" (UniqueName: \"kubernetes.io/projected/c48f3728-76d9-43c5-8053-b27ac3751ecb-kube-api-access-st8d5\") pod \"placement-962c-account-create-update-nbrwz\" (UID: \"c48f3728-76d9-43c5-8053-b27ac3751ecb\") " pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.629829 4936 generic.go:334] "Generic (PLEG): container finished" podID="223bcfee-4f1a-4ce5-9e38-a8bd4676411b" containerID="59a9b9149a96aa019822cbac94464d9c9c9b540e526b2838ca44aac6399da5fd" exitCode=0 Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.629913 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-22blq" event={"ID":"223bcfee-4f1a-4ce5-9e38-a8bd4676411b","Type":"ContainerDied","Data":"59a9b9149a96aa019822cbac94464d9c9c9b540e526b2838ca44aac6399da5fd"} Mar 20 16:17:52 crc kubenswrapper[4936]: I0320 16:17:52.694970 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:53 crc kubenswrapper[4936]: W0320 16:17:53.018745 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4f09dd_f403_4f3b_ac44_129df52fd796.slice/crio-6fd3cd409e55e50e1b2a7ee3055076f3edc72852e1f168bd2975a43939347333 WatchSource:0}: Error finding container 6fd3cd409e55e50e1b2a7ee3055076f3edc72852e1f168bd2975a43939347333: Status 404 returned error can't find the container with id 6fd3cd409e55e50e1b2a7ee3055076f3edc72852e1f168bd2975a43939347333 Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.035019 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.042121 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.210760 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-dns-svc\") pod \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.210937 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcp5t\" (UniqueName: \"kubernetes.io/projected/cd16e0f6-1913-4750-adad-8db14435dd01-kube-api-access-dcp5t\") pod \"cd16e0f6-1913-4750-adad-8db14435dd01\" (UID: \"cd16e0f6-1913-4750-adad-8db14435dd01\") " Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.210968 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-ovsdbserver-nb\") pod \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.211214 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znxsc\" (UniqueName: \"kubernetes.io/projected/f377f06a-3265-4b2a-8a33-1e39a572ecb0-kube-api-access-znxsc\") pod \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.211257 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-config\") pod \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\" (UID: \"f377f06a-3265-4b2a-8a33-1e39a572ecb0\") " Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.211284 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd16e0f6-1913-4750-adad-8db14435dd01-operator-scripts\") pod \"cd16e0f6-1913-4750-adad-8db14435dd01\" (UID: \"cd16e0f6-1913-4750-adad-8db14435dd01\") " Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.213316 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd16e0f6-1913-4750-adad-8db14435dd01-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cd16e0f6-1913-4750-adad-8db14435dd01" (UID: "cd16e0f6-1913-4750-adad-8db14435dd01"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.214098 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd16e0f6-1913-4750-adad-8db14435dd01-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.220140 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f377f06a-3265-4b2a-8a33-1e39a572ecb0-kube-api-access-znxsc" (OuterVolumeSpecName: "kube-api-access-znxsc") pod "f377f06a-3265-4b2a-8a33-1e39a572ecb0" (UID: "f377f06a-3265-4b2a-8a33-1e39a572ecb0"). InnerVolumeSpecName "kube-api-access-znxsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.221748 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd16e0f6-1913-4750-adad-8db14435dd01-kube-api-access-dcp5t" (OuterVolumeSpecName: "kube-api-access-dcp5t") pod "cd16e0f6-1913-4750-adad-8db14435dd01" (UID: "cd16e0f6-1913-4750-adad-8db14435dd01"). InnerVolumeSpecName "kube-api-access-dcp5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.230327 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f377f06a-3265-4b2a-8a33-1e39a572ecb0" (UID: "f377f06a-3265-4b2a-8a33-1e39a572ecb0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.234532 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-config" (OuterVolumeSpecName: "config") pod "f377f06a-3265-4b2a-8a33-1e39a572ecb0" (UID: "f377f06a-3265-4b2a-8a33-1e39a572ecb0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.240304 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f377f06a-3265-4b2a-8a33-1e39a572ecb0" (UID: "f377f06a-3265-4b2a-8a33-1e39a572ecb0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.315977 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.316012 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcp5t\" (UniqueName: \"kubernetes.io/projected/cd16e0f6-1913-4750-adad-8db14435dd01-kube-api-access-dcp5t\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.316025 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.316034 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znxsc\" (UniqueName: \"kubernetes.io/projected/f377f06a-3265-4b2a-8a33-1e39a572ecb0-kube-api-access-znxsc\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.316043 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f377f06a-3265-4b2a-8a33-1e39a572ecb0-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.642123 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" event={"ID":"f377f06a-3265-4b2a-8a33-1e39a572ecb0","Type":"ContainerDied","Data":"09579b62ab8fe01c9affe9f42584b7645487b92d8cf87c50174ed6c1f3b91e12"} Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.642138 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-ggvg7" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.642185 4936 scope.go:117] "RemoveContainer" containerID="4937fd5d5ca10717ae68151bac3416a90dc4f6e937df093f6a4e2ee952bd6ece" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.645500 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zv6gp" event={"ID":"cd16e0f6-1913-4750-adad-8db14435dd01","Type":"ContainerDied","Data":"4572eb60641e39587be859e399f036ed489cab2de2d4457653df779ef2febf51"} Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.645524 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4572eb60641e39587be859e399f036ed489cab2de2d4457653df779ef2febf51" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.645592 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zv6gp" Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.647242 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7f4f09dd-f403-4f3b-ac44-129df52fd796","Type":"ContainerStarted","Data":"6fd3cd409e55e50e1b2a7ee3055076f3edc72852e1f168bd2975a43939347333"} Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.701667 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-ggvg7"] Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.710116 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-ggvg7"] Mar 20 16:17:53 crc kubenswrapper[4936]: I0320 16:17:53.866786 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f377f06a-3265-4b2a-8a33-1e39a572ecb0" path="/var/lib/kubelet/pods/f377f06a-3265-4b2a-8a33-1e39a572ecb0/volumes" Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.670289 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-22blq" event={"ID":"223bcfee-4f1a-4ce5-9e38-a8bd4676411b","Type":"ContainerStarted","Data":"7cec49d48d8b5e1fcfe1151b1cb1485df8291075bb1345b9aa4d79fa5e372060"} Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.671038 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.674861 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6k8ph" event={"ID":"070d6439-9bd3-4fd8-b61a-df1f03badb79","Type":"ContainerStarted","Data":"8dc410f346ffc88c866c0183f21b8f06b51a3c1798e435a2867b867d56084513"} Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.698076 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9vpbv"] Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.707406 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-22blq" podStartSLOduration=4.707382778 podStartE2EDuration="4.707382778s" podCreationTimestamp="2026-03-20 16:17:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:17:54.688119596 +0000 UTC m=+1025.634487411" watchObservedRunningTime="2026-03-20 16:17:54.707382778 +0000 UTC m=+1025.653750593" Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.726727 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-6k8ph" podStartSLOduration=1.867799793 podStartE2EDuration="6.726700341s" podCreationTimestamp="2026-03-20 16:17:48 +0000 UTC" firstStartedPulling="2026-03-20 16:17:49.468891338 +0000 UTC m=+1020.415259153" lastFinishedPulling="2026-03-20 16:17:54.327791886 +0000 UTC m=+1025.274159701" observedRunningTime="2026-03-20 16:17:54.702693151 +0000 UTC m=+1025.649060976" watchObservedRunningTime="2026-03-20 16:17:54.726700341 +0000 UTC m=+1025.673068166" Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.818128 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-962c-account-create-update-nbrwz"] Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.825270 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jvq45"] Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.834021 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zphvf"] Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.841149 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7948-account-create-update-8trc2"] Mar 20 16:17:54 crc kubenswrapper[4936]: W0320 16:17:54.933331 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc48f3728_76d9_43c5_8053_b27ac3751ecb.slice/crio-7ed5d54df0e68a3dd5c8d9df4712649c64bf7dcaf86e809c4656195e8e532933 WatchSource:0}: Error finding container 7ed5d54df0e68a3dd5c8d9df4712649c64bf7dcaf86e809c4656195e8e532933: Status 404 returned error can't find the container with id 7ed5d54df0e68a3dd5c8d9df4712649c64bf7dcaf86e809c4656195e8e532933 Mar 20 16:17:54 crc kubenswrapper[4936]: W0320 16:17:54.934841 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod434fff7b_3810_45b8_8028_98a9d77d3f15.slice/crio-d2fe3f2d8639eba1bf32383d65e14790950327877184a3414fa8c233452bf58a WatchSource:0}: Error finding container d2fe3f2d8639eba1bf32383d65e14790950327877184a3414fa8c233452bf58a: Status 404 returned error can't find the container with id d2fe3f2d8639eba1bf32383d65e14790950327877184a3414fa8c233452bf58a Mar 20 16:17:54 crc kubenswrapper[4936]: W0320 16:17:54.936516 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod812b9782_84b1_41fc_939e_30e82c1b5e13.slice/crio-b1627ef7c52e522d530e55435481d75fb6c5a57ea402e077003714c405f3ded8 WatchSource:0}: Error finding container b1627ef7c52e522d530e55435481d75fb6c5a57ea402e077003714c405f3ded8: Status 404 returned error can't find the container with id b1627ef7c52e522d530e55435481d75fb6c5a57ea402e077003714c405f3ded8 Mar 20 16:17:54 crc kubenswrapper[4936]: W0320 16:17:54.937755 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8df2114_9d0b_43c8_91cd_a60355f7238b.slice/crio-a00b37be3ff133298a0c9875e61262f03bd844ce218aaab9803362934e9c1009 WatchSource:0}: Error finding container a00b37be3ff133298a0c9875e61262f03bd844ce218aaab9803362934e9c1009: Status 404 returned error can't find the container with id a00b37be3ff133298a0c9875e61262f03bd844ce218aaab9803362934e9c1009 Mar 20 16:17:54 crc kubenswrapper[4936]: W0320 16:17:54.940303 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0aa18c2_d52b_4671_996b_985c848a748f.slice/crio-8c5b4fe70fdd58ed26f281b7eb07147a076d4033169952c3a99045d22232f617 WatchSource:0}: Error finding container 8c5b4fe70fdd58ed26f281b7eb07147a076d4033169952c3a99045d22232f617: Status 404 returned error can't find the container with id 8c5b4fe70fdd58ed26f281b7eb07147a076d4033169952c3a99045d22232f617 Mar 20 16:17:54 crc kubenswrapper[4936]: I0320 16:17:54.993141 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-05f5-account-create-update-9vd8r"] Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.705398 4936 generic.go:334] "Generic (PLEG): container finished" podID="c48f3728-76d9-43c5-8053-b27ac3751ecb" containerID="f52edbe07ed941a85bc6a10d310a58fcaeb3c1ad929a5efa5f9ee8810ad218f9" exitCode=0 Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.705580 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-962c-account-create-update-nbrwz" event={"ID":"c48f3728-76d9-43c5-8053-b27ac3751ecb","Type":"ContainerDied","Data":"f52edbe07ed941a85bc6a10d310a58fcaeb3c1ad929a5efa5f9ee8810ad218f9"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.705782 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-962c-account-create-update-nbrwz" event={"ID":"c48f3728-76d9-43c5-8053-b27ac3751ecb","Type":"ContainerStarted","Data":"7ed5d54df0e68a3dd5c8d9df4712649c64bf7dcaf86e809c4656195e8e532933"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.707979 4936 generic.go:334] "Generic (PLEG): container finished" podID="434fff7b-3810-45b8-8028-98a9d77d3f15" containerID="75f5f9b53ac503449b410ff96fd3effa3da09f9333ba2c1c601660b6f5a88afb" exitCode=0 Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.708318 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zphvf" event={"ID":"434fff7b-3810-45b8-8028-98a9d77d3f15","Type":"ContainerDied","Data":"75f5f9b53ac503449b410ff96fd3effa3da09f9333ba2c1c601660b6f5a88afb"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.708346 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zphvf" event={"ID":"434fff7b-3810-45b8-8028-98a9d77d3f15","Type":"ContainerStarted","Data":"d2fe3f2d8639eba1bf32383d65e14790950327877184a3414fa8c233452bf58a"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.714329 4936 generic.go:334] "Generic (PLEG): container finished" podID="a0aa18c2-d52b-4671-996b-985c848a748f" containerID="666a52a4fc6ae392ee80e61155687b982c432fe0f44003898f24d68084097b76" exitCode=0 Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.714442 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7948-account-create-update-8trc2" event={"ID":"a0aa18c2-d52b-4671-996b-985c848a748f","Type":"ContainerDied","Data":"666a52a4fc6ae392ee80e61155687b982c432fe0f44003898f24d68084097b76"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.714491 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7948-account-create-update-8trc2" event={"ID":"a0aa18c2-d52b-4671-996b-985c848a748f","Type":"ContainerStarted","Data":"8c5b4fe70fdd58ed26f281b7eb07147a076d4033169952c3a99045d22232f617"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.716593 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7f4f09dd-f403-4f3b-ac44-129df52fd796","Type":"ContainerStarted","Data":"fd019512e526412267195dcc6d31120c6cf0b2e7d63441d37182d617bf402014"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.718321 4936 generic.go:334] "Generic (PLEG): container finished" podID="812b9782-84b1-41fc-939e-30e82c1b5e13" containerID="2f49546a1d1fa1324c648ba58556f65bb5b210cdb9ce4ff81c0f4e813897788d" exitCode=0 Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.718362 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9vpbv" event={"ID":"812b9782-84b1-41fc-939e-30e82c1b5e13","Type":"ContainerDied","Data":"2f49546a1d1fa1324c648ba58556f65bb5b210cdb9ce4ff81c0f4e813897788d"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.718376 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9vpbv" event={"ID":"812b9782-84b1-41fc-939e-30e82c1b5e13","Type":"ContainerStarted","Data":"b1627ef7c52e522d530e55435481d75fb6c5a57ea402e077003714c405f3ded8"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.720090 4936 generic.go:334] "Generic (PLEG): container finished" podID="b8df2114-9d0b-43c8-91cd-a60355f7238b" containerID="2d73a47e2620619edcff2f7a567e03b95fd0efb034c602109094656476739fc1" exitCode=0 Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.720230 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jvq45" event={"ID":"b8df2114-9d0b-43c8-91cd-a60355f7238b","Type":"ContainerDied","Data":"2d73a47e2620619edcff2f7a567e03b95fd0efb034c602109094656476739fc1"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.720246 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jvq45" event={"ID":"b8df2114-9d0b-43c8-91cd-a60355f7238b","Type":"ContainerStarted","Data":"a00b37be3ff133298a0c9875e61262f03bd844ce218aaab9803362934e9c1009"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.729007 4936 generic.go:334] "Generic (PLEG): container finished" podID="11a87bf1-c0b8-47a2-8ccb-faf322bebfb1" containerID="74158cac144ee91c9dfde123deaa764b39ca9122d5eaaf455154f644623db967" exitCode=0 Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.729318 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-05f5-account-create-update-9vd8r" event={"ID":"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1","Type":"ContainerDied","Data":"74158cac144ee91c9dfde123deaa764b39ca9122d5eaaf455154f644623db967"} Mar 20 16:17:55 crc kubenswrapper[4936]: I0320 16:17:55.729358 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-05f5-account-create-update-9vd8r" event={"ID":"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1","Type":"ContainerStarted","Data":"0271eb8cc5d3b7d72a40b3fb7959272b02b3f3d517c3bd71c4d479e7f01c9ff7"} Mar 20 16:17:56 crc kubenswrapper[4936]: I0320 16:17:56.739863 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7f4f09dd-f403-4f3b-ac44-129df52fd796","Type":"ContainerStarted","Data":"985498a2c769327b3ec5711f2f2abb201ae1541b2a0a756f627c319e7831a0c3"} Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.152963 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.182006 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=5.205904518 podStartE2EDuration="7.181987184s" podCreationTimestamp="2026-03-20 16:17:50 +0000 UTC" firstStartedPulling="2026-03-20 16:17:53.020214781 +0000 UTC m=+1023.966582606" lastFinishedPulling="2026-03-20 16:17:54.996297457 +0000 UTC m=+1025.942665272" observedRunningTime="2026-03-20 16:17:56.771610004 +0000 UTC m=+1027.717977819" watchObservedRunningTime="2026-03-20 16:17:57.181987184 +0000 UTC m=+1028.128354999" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.288517 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxxmw\" (UniqueName: \"kubernetes.io/projected/812b9782-84b1-41fc-939e-30e82c1b5e13-kube-api-access-jxxmw\") pod \"812b9782-84b1-41fc-939e-30e82c1b5e13\" (UID: \"812b9782-84b1-41fc-939e-30e82c1b5e13\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.288814 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/812b9782-84b1-41fc-939e-30e82c1b5e13-operator-scripts\") pod \"812b9782-84b1-41fc-939e-30e82c1b5e13\" (UID: \"812b9782-84b1-41fc-939e-30e82c1b5e13\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.289876 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/812b9782-84b1-41fc-939e-30e82c1b5e13-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "812b9782-84b1-41fc-939e-30e82c1b5e13" (UID: "812b9782-84b1-41fc-939e-30e82c1b5e13"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.299551 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/812b9782-84b1-41fc-939e-30e82c1b5e13-kube-api-access-jxxmw" (OuterVolumeSpecName: "kube-api-access-jxxmw") pod "812b9782-84b1-41fc-939e-30e82c1b5e13" (UID: "812b9782-84b1-41fc-939e-30e82c1b5e13"). InnerVolumeSpecName "kube-api-access-jxxmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.390880 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/812b9782-84b1-41fc-939e-30e82c1b5e13-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.390930 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxxmw\" (UniqueName: \"kubernetes.io/projected/812b9782-84b1-41fc-939e-30e82c1b5e13-kube-api-access-jxxmw\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.413689 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.421761 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.483584 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.492180 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2v8h\" (UniqueName: \"kubernetes.io/projected/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-kube-api-access-n2v8h\") pod \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\" (UID: \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.492273 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-operator-scripts\") pod \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\" (UID: \"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.492413 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xccl\" (UniqueName: \"kubernetes.io/projected/434fff7b-3810-45b8-8028-98a9d77d3f15-kube-api-access-8xccl\") pod \"434fff7b-3810-45b8-8028-98a9d77d3f15\" (UID: \"434fff7b-3810-45b8-8028-98a9d77d3f15\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.492486 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/434fff7b-3810-45b8-8028-98a9d77d3f15-operator-scripts\") pod \"434fff7b-3810-45b8-8028-98a9d77d3f15\" (UID: \"434fff7b-3810-45b8-8028-98a9d77d3f15\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.493488 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/434fff7b-3810-45b8-8028-98a9d77d3f15-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "434fff7b-3810-45b8-8028-98a9d77d3f15" (UID: "434fff7b-3810-45b8-8028-98a9d77d3f15"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.494757 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "11a87bf1-c0b8-47a2-8ccb-faf322bebfb1" (UID: "11a87bf1-c0b8-47a2-8ccb-faf322bebfb1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.497013 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jvq45" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.498653 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-kube-api-access-n2v8h" (OuterVolumeSpecName: "kube-api-access-n2v8h") pod "11a87bf1-c0b8-47a2-8ccb-faf322bebfb1" (UID: "11a87bf1-c0b8-47a2-8ccb-faf322bebfb1"). InnerVolumeSpecName "kube-api-access-n2v8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.529081 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434fff7b-3810-45b8-8028-98a9d77d3f15-kube-api-access-8xccl" (OuterVolumeSpecName: "kube-api-access-8xccl") pod "434fff7b-3810-45b8-8028-98a9d77d3f15" (UID: "434fff7b-3810-45b8-8028-98a9d77d3f15"). InnerVolumeSpecName "kube-api-access-8xccl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.571626 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.593680 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8df2114-9d0b-43c8-91cd-a60355f7238b-operator-scripts\") pod \"b8df2114-9d0b-43c8-91cd-a60355f7238b\" (UID: \"b8df2114-9d0b-43c8-91cd-a60355f7238b\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.593756 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqrh9\" (UniqueName: \"kubernetes.io/projected/b8df2114-9d0b-43c8-91cd-a60355f7238b-kube-api-access-xqrh9\") pod \"b8df2114-9d0b-43c8-91cd-a60355f7238b\" (UID: \"b8df2114-9d0b-43c8-91cd-a60355f7238b\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.593826 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st8d5\" (UniqueName: \"kubernetes.io/projected/c48f3728-76d9-43c5-8053-b27ac3751ecb-kube-api-access-st8d5\") pod \"c48f3728-76d9-43c5-8053-b27ac3751ecb\" (UID: \"c48f3728-76d9-43c5-8053-b27ac3751ecb\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.593965 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48f3728-76d9-43c5-8053-b27ac3751ecb-operator-scripts\") pod \"c48f3728-76d9-43c5-8053-b27ac3751ecb\" (UID: \"c48f3728-76d9-43c5-8053-b27ac3751ecb\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.594278 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8df2114-9d0b-43c8-91cd-a60355f7238b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b8df2114-9d0b-43c8-91cd-a60355f7238b" (UID: "b8df2114-9d0b-43c8-91cd-a60355f7238b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.594357 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xccl\" (UniqueName: \"kubernetes.io/projected/434fff7b-3810-45b8-8028-98a9d77d3f15-kube-api-access-8xccl\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.594373 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/434fff7b-3810-45b8-8028-98a9d77d3f15-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.594382 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2v8h\" (UniqueName: \"kubernetes.io/projected/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-kube-api-access-n2v8h\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.594393 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.594656 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48f3728-76d9-43c5-8053-b27ac3751ecb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c48f3728-76d9-43c5-8053-b27ac3751ecb" (UID: "c48f3728-76d9-43c5-8053-b27ac3751ecb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.597455 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48f3728-76d9-43c5-8053-b27ac3751ecb-kube-api-access-st8d5" (OuterVolumeSpecName: "kube-api-access-st8d5") pod "c48f3728-76d9-43c5-8053-b27ac3751ecb" (UID: "c48f3728-76d9-43c5-8053-b27ac3751ecb"). InnerVolumeSpecName "kube-api-access-st8d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.599688 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8df2114-9d0b-43c8-91cd-a60355f7238b-kube-api-access-xqrh9" (OuterVolumeSpecName: "kube-api-access-xqrh9") pod "b8df2114-9d0b-43c8-91cd-a60355f7238b" (UID: "b8df2114-9d0b-43c8-91cd-a60355f7238b"). InnerVolumeSpecName "kube-api-access-xqrh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.697418 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0aa18c2-d52b-4671-996b-985c848a748f-operator-scripts\") pod \"a0aa18c2-d52b-4671-996b-985c848a748f\" (UID: \"a0aa18c2-d52b-4671-996b-985c848a748f\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.698865 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr4jb\" (UniqueName: \"kubernetes.io/projected/a0aa18c2-d52b-4671-996b-985c848a748f-kube-api-access-rr4jb\") pod \"a0aa18c2-d52b-4671-996b-985c848a748f\" (UID: \"a0aa18c2-d52b-4671-996b-985c848a748f\") " Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.698229 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0aa18c2-d52b-4671-996b-985c848a748f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a0aa18c2-d52b-4671-996b-985c848a748f" (UID: "a0aa18c2-d52b-4671-996b-985c848a748f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.699867 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st8d5\" (UniqueName: \"kubernetes.io/projected/c48f3728-76d9-43c5-8053-b27ac3751ecb-kube-api-access-st8d5\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.699995 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48f3728-76d9-43c5-8053-b27ac3751ecb-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.700082 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8df2114-9d0b-43c8-91cd-a60355f7238b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.700163 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqrh9\" (UniqueName: \"kubernetes.io/projected/b8df2114-9d0b-43c8-91cd-a60355f7238b-kube-api-access-xqrh9\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.700254 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0aa18c2-d52b-4671-996b-985c848a748f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.702471 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0aa18c2-d52b-4671-996b-985c848a748f-kube-api-access-rr4jb" (OuterVolumeSpecName: "kube-api-access-rr4jb") pod "a0aa18c2-d52b-4671-996b-985c848a748f" (UID: "a0aa18c2-d52b-4671-996b-985c848a748f"). InnerVolumeSpecName "kube-api-access-rr4jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.748075 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-05f5-account-create-update-9vd8r" event={"ID":"11a87bf1-c0b8-47a2-8ccb-faf322bebfb1","Type":"ContainerDied","Data":"0271eb8cc5d3b7d72a40b3fb7959272b02b3f3d517c3bd71c4d479e7f01c9ff7"} Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.748110 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0271eb8cc5d3b7d72a40b3fb7959272b02b3f3d517c3bd71c4d479e7f01c9ff7" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.748164 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-05f5-account-create-update-9vd8r" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.750652 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-962c-account-create-update-nbrwz" event={"ID":"c48f3728-76d9-43c5-8053-b27ac3751ecb","Type":"ContainerDied","Data":"7ed5d54df0e68a3dd5c8d9df4712649c64bf7dcaf86e809c4656195e8e532933"} Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.750686 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ed5d54df0e68a3dd5c8d9df4712649c64bf7dcaf86e809c4656195e8e532933" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.750743 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-962c-account-create-update-nbrwz" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.753463 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zphvf" event={"ID":"434fff7b-3810-45b8-8028-98a9d77d3f15","Type":"ContainerDied","Data":"d2fe3f2d8639eba1bf32383d65e14790950327877184a3414fa8c233452bf58a"} Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.753503 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2fe3f2d8639eba1bf32383d65e14790950327877184a3414fa8c233452bf58a" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.753580 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zphvf" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.757126 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7948-account-create-update-8trc2" event={"ID":"a0aa18c2-d52b-4671-996b-985c848a748f","Type":"ContainerDied","Data":"8c5b4fe70fdd58ed26f281b7eb07147a076d4033169952c3a99045d22232f617"} Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.757159 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7948-account-create-update-8trc2" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.757161 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c5b4fe70fdd58ed26f281b7eb07147a076d4033169952c3a99045d22232f617" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.758634 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9vpbv" event={"ID":"812b9782-84b1-41fc-939e-30e82c1b5e13","Type":"ContainerDied","Data":"b1627ef7c52e522d530e55435481d75fb6c5a57ea402e077003714c405f3ded8"} Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.758663 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1627ef7c52e522d530e55435481d75fb6c5a57ea402e077003714c405f3ded8" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.758754 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9vpbv" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.760205 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jvq45" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.760186 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jvq45" event={"ID":"b8df2114-9d0b-43c8-91cd-a60355f7238b","Type":"ContainerDied","Data":"a00b37be3ff133298a0c9875e61262f03bd844ce218aaab9803362934e9c1009"} Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.760246 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a00b37be3ff133298a0c9875e61262f03bd844ce218aaab9803362934e9c1009" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.760486 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 20 16:17:57 crc kubenswrapper[4936]: I0320 16:17:57.802296 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr4jb\" (UniqueName: \"kubernetes.io/projected/a0aa18c2-d52b-4671-996b-985c848a748f-kube-api-access-rr4jb\") on node \"crc\" DevicePath \"\"" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.069279 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-zv6gp"] Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.075203 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-zv6gp"] Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.162486 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-cjk9m"] Mar 20 16:17:58 crc kubenswrapper[4936]: E0320 16:17:58.162958 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f377f06a-3265-4b2a-8a33-1e39a572ecb0" containerName="init" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.162978 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f377f06a-3265-4b2a-8a33-1e39a572ecb0" containerName="init" Mar 20 16:17:58 crc kubenswrapper[4936]: E0320 16:17:58.163007 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a87bf1-c0b8-47a2-8ccb-faf322bebfb1" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163016 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a87bf1-c0b8-47a2-8ccb-faf322bebfb1" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: E0320 16:17:58.163031 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8df2114-9d0b-43c8-91cd-a60355f7238b" containerName="mariadb-database-create" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163040 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8df2114-9d0b-43c8-91cd-a60355f7238b" containerName="mariadb-database-create" Mar 20 16:17:58 crc kubenswrapper[4936]: E0320 16:17:58.163050 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434fff7b-3810-45b8-8028-98a9d77d3f15" containerName="mariadb-database-create" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163058 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="434fff7b-3810-45b8-8028-98a9d77d3f15" containerName="mariadb-database-create" Mar 20 16:17:58 crc kubenswrapper[4936]: E0320 16:17:58.163074 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48f3728-76d9-43c5-8053-b27ac3751ecb" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163083 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48f3728-76d9-43c5-8053-b27ac3751ecb" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: E0320 16:17:58.163101 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0aa18c2-d52b-4671-996b-985c848a748f" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163109 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0aa18c2-d52b-4671-996b-985c848a748f" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: E0320 16:17:58.163126 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd16e0f6-1913-4750-adad-8db14435dd01" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163134 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd16e0f6-1913-4750-adad-8db14435dd01" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: E0320 16:17:58.163149 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812b9782-84b1-41fc-939e-30e82c1b5e13" containerName="mariadb-database-create" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163157 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="812b9782-84b1-41fc-939e-30e82c1b5e13" containerName="mariadb-database-create" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163354 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0aa18c2-d52b-4671-996b-985c848a748f" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163375 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="c48f3728-76d9-43c5-8053-b27ac3751ecb" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163384 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="434fff7b-3810-45b8-8028-98a9d77d3f15" containerName="mariadb-database-create" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163394 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="11a87bf1-c0b8-47a2-8ccb-faf322bebfb1" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163408 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f377f06a-3265-4b2a-8a33-1e39a572ecb0" containerName="init" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163420 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd16e0f6-1913-4750-adad-8db14435dd01" containerName="mariadb-account-create-update" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163428 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="812b9782-84b1-41fc-939e-30e82c1b5e13" containerName="mariadb-database-create" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.163440 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8df2114-9d0b-43c8-91cd-a60355f7238b" containerName="mariadb-database-create" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.164132 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cjk9m" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.170800 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-cjk9m"] Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.174092 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.310526 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-operator-scripts\") pod \"root-account-create-update-cjk9m\" (UID: \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\") " pod="openstack/root-account-create-update-cjk9m" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.310631 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpxtf\" (UniqueName: \"kubernetes.io/projected/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-kube-api-access-zpxtf\") pod \"root-account-create-update-cjk9m\" (UID: \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\") " pod="openstack/root-account-create-update-cjk9m" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.412122 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-operator-scripts\") pod \"root-account-create-update-cjk9m\" (UID: \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\") " pod="openstack/root-account-create-update-cjk9m" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.412530 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpxtf\" (UniqueName: \"kubernetes.io/projected/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-kube-api-access-zpxtf\") pod \"root-account-create-update-cjk9m\" (UID: \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\") " pod="openstack/root-account-create-update-cjk9m" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.412990 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-operator-scripts\") pod \"root-account-create-update-cjk9m\" (UID: \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\") " pod="openstack/root-account-create-update-cjk9m" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.429809 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpxtf\" (UniqueName: \"kubernetes.io/projected/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-kube-api-access-zpxtf\") pod \"root-account-create-update-cjk9m\" (UID: \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\") " pod="openstack/root-account-create-update-cjk9m" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.484702 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cjk9m" Mar 20 16:17:58 crc kubenswrapper[4936]: I0320 16:17:58.936717 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-cjk9m"] Mar 20 16:17:59 crc kubenswrapper[4936]: I0320 16:17:59.777500 4936 generic.go:334] "Generic (PLEG): container finished" podID="3aece10e-0797-48c2-aac7-7a4b12cf0cf5" containerID="9d1d5fa5c224fda226d2cb9276b7341cf1d644f1375cb62c6e65bddb1a799120" exitCode=0 Mar 20 16:17:59 crc kubenswrapper[4936]: I0320 16:17:59.777637 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cjk9m" event={"ID":"3aece10e-0797-48c2-aac7-7a4b12cf0cf5","Type":"ContainerDied","Data":"9d1d5fa5c224fda226d2cb9276b7341cf1d644f1375cb62c6e65bddb1a799120"} Mar 20 16:17:59 crc kubenswrapper[4936]: I0320 16:17:59.777872 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cjk9m" event={"ID":"3aece10e-0797-48c2-aac7-7a4b12cf0cf5","Type":"ContainerStarted","Data":"55bb86347f9740e04ba956167fd81d92bc254a1b6e5f35df92b76d52b6ad62f4"} Mar 20 16:17:59 crc kubenswrapper[4936]: I0320 16:17:59.876307 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd16e0f6-1913-4750-adad-8db14435dd01" path="/var/lib/kubelet/pods/cd16e0f6-1913-4750-adad-8db14435dd01/volumes" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.135720 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567058-c652h"] Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.137523 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567058-c652h" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.142563 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.142891 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.143120 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.145078 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567058-c652h"] Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.243302 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vcx7\" (UniqueName: \"kubernetes.io/projected/d274515c-4ea9-459b-b934-ffd2d6982d5c-kube-api-access-8vcx7\") pod \"auto-csr-approver-29567058-c652h\" (UID: \"d274515c-4ea9-459b-b934-ffd2d6982d5c\") " pod="openshift-infra/auto-csr-approver-29567058-c652h" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.344745 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vcx7\" (UniqueName: \"kubernetes.io/projected/d274515c-4ea9-459b-b934-ffd2d6982d5c-kube-api-access-8vcx7\") pod \"auto-csr-approver-29567058-c652h\" (UID: \"d274515c-4ea9-459b-b934-ffd2d6982d5c\") " pod="openshift-infra/auto-csr-approver-29567058-c652h" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.364526 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vcx7\" (UniqueName: \"kubernetes.io/projected/d274515c-4ea9-459b-b934-ffd2d6982d5c-kube-api-access-8vcx7\") pod \"auto-csr-approver-29567058-c652h\" (UID: \"d274515c-4ea9-459b-b934-ffd2d6982d5c\") " pod="openshift-infra/auto-csr-approver-29567058-c652h" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.455384 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567058-c652h" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.547369 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:18:00 crc kubenswrapper[4936]: E0320 16:18:00.547618 4936 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 20 16:18:00 crc kubenswrapper[4936]: E0320 16:18:00.547639 4936 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 20 16:18:00 crc kubenswrapper[4936]: E0320 16:18:00.547688 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift podName:b1ea233b-41e7-49ac-a878-aa038b350252 nodeName:}" failed. No retries permitted until 2026-03-20 16:18:16.547668576 +0000 UTC m=+1047.494036401 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift") pod "swift-storage-0" (UID: "b1ea233b-41e7-49ac-a878-aa038b350252") : configmap "swift-ring-files" not found Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.647830 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.719638 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-79vw6"] Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.719997 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" podUID="e353d140-f7e5-47b4-b18e-200d5a255086" containerName="dnsmasq-dns" containerID="cri-o://8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d" gracePeriod=10 Mar 20 16:18:00 crc kubenswrapper[4936]: I0320 16:18:00.904070 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567058-c652h"] Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.158930 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cjk9m" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.261557 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-operator-scripts\") pod \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\" (UID: \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\") " Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.261882 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpxtf\" (UniqueName: \"kubernetes.io/projected/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-kube-api-access-zpxtf\") pod \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\" (UID: \"3aece10e-0797-48c2-aac7-7a4b12cf0cf5\") " Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.262220 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3aece10e-0797-48c2-aac7-7a4b12cf0cf5" (UID: "3aece10e-0797-48c2-aac7-7a4b12cf0cf5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.267613 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-kube-api-access-zpxtf" (OuterVolumeSpecName: "kube-api-access-zpxtf") pod "3aece10e-0797-48c2-aac7-7a4b12cf0cf5" (UID: "3aece10e-0797-48c2-aac7-7a4b12cf0cf5"). InnerVolumeSpecName "kube-api-access-zpxtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.285595 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.362709 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-dns-svc\") pod \"e353d140-f7e5-47b4-b18e-200d5a255086\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.362847 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kxxh\" (UniqueName: \"kubernetes.io/projected/e353d140-f7e5-47b4-b18e-200d5a255086-kube-api-access-8kxxh\") pod \"e353d140-f7e5-47b4-b18e-200d5a255086\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.362935 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-config\") pod \"e353d140-f7e5-47b4-b18e-200d5a255086\" (UID: \"e353d140-f7e5-47b4-b18e-200d5a255086\") " Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.363352 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpxtf\" (UniqueName: \"kubernetes.io/projected/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-kube-api-access-zpxtf\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.363372 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aece10e-0797-48c2-aac7-7a4b12cf0cf5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.367037 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e353d140-f7e5-47b4-b18e-200d5a255086-kube-api-access-8kxxh" (OuterVolumeSpecName: "kube-api-access-8kxxh") pod "e353d140-f7e5-47b4-b18e-200d5a255086" (UID: "e353d140-f7e5-47b4-b18e-200d5a255086"). InnerVolumeSpecName "kube-api-access-8kxxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.397125 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-config" (OuterVolumeSpecName: "config") pod "e353d140-f7e5-47b4-b18e-200d5a255086" (UID: "e353d140-f7e5-47b4-b18e-200d5a255086"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.403175 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e353d140-f7e5-47b4-b18e-200d5a255086" (UID: "e353d140-f7e5-47b4-b18e-200d5a255086"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.465394 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kxxh\" (UniqueName: \"kubernetes.io/projected/e353d140-f7e5-47b4-b18e-200d5a255086-kube-api-access-8kxxh\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.465438 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.465448 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e353d140-f7e5-47b4-b18e-200d5a255086-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.795854 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567058-c652h" event={"ID":"d274515c-4ea9-459b-b934-ffd2d6982d5c","Type":"ContainerStarted","Data":"f60b06654f5992cfd15834b2fffd483cb501a2737ebd29dfb7ba20d3318ef014"} Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.797658 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cjk9m" event={"ID":"3aece10e-0797-48c2-aac7-7a4b12cf0cf5","Type":"ContainerDied","Data":"55bb86347f9740e04ba956167fd81d92bc254a1b6e5f35df92b76d52b6ad62f4"} Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.797726 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55bb86347f9740e04ba956167fd81d92bc254a1b6e5f35df92b76d52b6ad62f4" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.797694 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cjk9m" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.799693 4936 generic.go:334] "Generic (PLEG): container finished" podID="e353d140-f7e5-47b4-b18e-200d5a255086" containerID="8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d" exitCode=0 Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.799739 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.799752 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" event={"ID":"e353d140-f7e5-47b4-b18e-200d5a255086","Type":"ContainerDied","Data":"8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d"} Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.799777 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-79vw6" event={"ID":"e353d140-f7e5-47b4-b18e-200d5a255086","Type":"ContainerDied","Data":"6de6e379cd2a7715ac771f6329de8a4e268bcd71bade8eab1d4dbb1f2452b1fb"} Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.799793 4936 scope.go:117] "RemoveContainer" containerID="8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.801668 4936 generic.go:334] "Generic (PLEG): container finished" podID="070d6439-9bd3-4fd8-b61a-df1f03badb79" containerID="8dc410f346ffc88c866c0183f21b8f06b51a3c1798e435a2867b867d56084513" exitCode=0 Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.801707 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6k8ph" event={"ID":"070d6439-9bd3-4fd8-b61a-df1f03badb79","Type":"ContainerDied","Data":"8dc410f346ffc88c866c0183f21b8f06b51a3c1798e435a2867b867d56084513"} Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.823793 4936 scope.go:117] "RemoveContainer" containerID="002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.842155 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-79vw6"] Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.853412 4936 scope.go:117] "RemoveContainer" containerID="8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d" Mar 20 16:18:01 crc kubenswrapper[4936]: E0320 16:18:01.855660 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d\": container with ID starting with 8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d not found: ID does not exist" containerID="8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.855720 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d"} err="failed to get container status \"8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d\": rpc error: code = NotFound desc = could not find container \"8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d\": container with ID starting with 8eb9b4cc0f80810a5aa2f21230c7af7ddb688c7c746041791edb6bad9ae6d60d not found: ID does not exist" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.855755 4936 scope.go:117] "RemoveContainer" containerID="002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c" Mar 20 16:18:01 crc kubenswrapper[4936]: E0320 16:18:01.859068 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c\": container with ID starting with 002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c not found: ID does not exist" containerID="002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.859151 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c"} err="failed to get container status \"002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c\": rpc error: code = NotFound desc = could not find container \"002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c\": container with ID starting with 002ce6d9884232452018482724c47932861362346e9cb7a7d04cd5cab673df6c not found: ID does not exist" Mar 20 16:18:01 crc kubenswrapper[4936]: I0320 16:18:01.872513 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-79vw6"] Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.087684 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-sg9df"] Mar 20 16:18:02 crc kubenswrapper[4936]: E0320 16:18:02.088620 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e353d140-f7e5-47b4-b18e-200d5a255086" containerName="dnsmasq-dns" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.088738 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="e353d140-f7e5-47b4-b18e-200d5a255086" containerName="dnsmasq-dns" Mar 20 16:18:02 crc kubenswrapper[4936]: E0320 16:18:02.088829 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e353d140-f7e5-47b4-b18e-200d5a255086" containerName="init" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.088900 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="e353d140-f7e5-47b4-b18e-200d5a255086" containerName="init" Mar 20 16:18:02 crc kubenswrapper[4936]: E0320 16:18:02.088964 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aece10e-0797-48c2-aac7-7a4b12cf0cf5" containerName="mariadb-account-create-update" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.089022 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aece10e-0797-48c2-aac7-7a4b12cf0cf5" containerName="mariadb-account-create-update" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.089264 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="e353d140-f7e5-47b4-b18e-200d5a255086" containerName="dnsmasq-dns" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.089359 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aece10e-0797-48c2-aac7-7a4b12cf0cf5" containerName="mariadb-account-create-update" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.090015 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.093063 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.093398 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-l2w6m" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.097204 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sg9df"] Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.175699 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-db-sync-config-data\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.175742 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-config-data\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.175776 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drsfw\" (UniqueName: \"kubernetes.io/projected/a817ffe2-32a4-4c37-9342-5d642d179ea7-kube-api-access-drsfw\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.175873 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-combined-ca-bundle\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.277172 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-combined-ca-bundle\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.277578 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-db-sync-config-data\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.277696 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-config-data\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.277811 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drsfw\" (UniqueName: \"kubernetes.io/projected/a817ffe2-32a4-4c37-9342-5d642d179ea7-kube-api-access-drsfw\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.283176 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-db-sync-config-data\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.283370 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-combined-ca-bundle\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.283704 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-config-data\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.298313 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drsfw\" (UniqueName: \"kubernetes.io/projected/a817ffe2-32a4-4c37-9342-5d642d179ea7-kube-api-access-drsfw\") pod \"glance-db-sync-sg9df\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.414943 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.811055 4936 generic.go:334] "Generic (PLEG): container finished" podID="d274515c-4ea9-459b-b934-ffd2d6982d5c" containerID="0ca535be5c5f888e070a3a4225ebb14c0abc608963f4c544653e596af6be8265" exitCode=0 Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.811105 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567058-c652h" event={"ID":"d274515c-4ea9-459b-b934-ffd2d6982d5c","Type":"ContainerDied","Data":"0ca535be5c5f888e070a3a4225ebb14c0abc608963f4c544653e596af6be8265"} Mar 20 16:18:02 crc kubenswrapper[4936]: I0320 16:18:02.985713 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sg9df"] Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.142925 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.200050 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-swiftconf\") pod \"070d6439-9bd3-4fd8-b61a-df1f03badb79\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.200122 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-ring-data-devices\") pod \"070d6439-9bd3-4fd8-b61a-df1f03badb79\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.200207 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/070d6439-9bd3-4fd8-b61a-df1f03badb79-etc-swift\") pod \"070d6439-9bd3-4fd8-b61a-df1f03badb79\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.200254 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-combined-ca-bundle\") pod \"070d6439-9bd3-4fd8-b61a-df1f03badb79\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.200283 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdvw9\" (UniqueName: \"kubernetes.io/projected/070d6439-9bd3-4fd8-b61a-df1f03badb79-kube-api-access-wdvw9\") pod \"070d6439-9bd3-4fd8-b61a-df1f03badb79\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.200401 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-scripts\") pod \"070d6439-9bd3-4fd8-b61a-df1f03badb79\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.200434 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-dispersionconf\") pod \"070d6439-9bd3-4fd8-b61a-df1f03badb79\" (UID: \"070d6439-9bd3-4fd8-b61a-df1f03badb79\") " Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.202213 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "070d6439-9bd3-4fd8-b61a-df1f03badb79" (UID: "070d6439-9bd3-4fd8-b61a-df1f03badb79"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.202650 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/070d6439-9bd3-4fd8-b61a-df1f03badb79-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "070d6439-9bd3-4fd8-b61a-df1f03badb79" (UID: "070d6439-9bd3-4fd8-b61a-df1f03badb79"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.208044 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/070d6439-9bd3-4fd8-b61a-df1f03badb79-kube-api-access-wdvw9" (OuterVolumeSpecName: "kube-api-access-wdvw9") pod "070d6439-9bd3-4fd8-b61a-df1f03badb79" (UID: "070d6439-9bd3-4fd8-b61a-df1f03badb79"). InnerVolumeSpecName "kube-api-access-wdvw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.208409 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "070d6439-9bd3-4fd8-b61a-df1f03badb79" (UID: "070d6439-9bd3-4fd8-b61a-df1f03badb79"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.224789 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-scripts" (OuterVolumeSpecName: "scripts") pod "070d6439-9bd3-4fd8-b61a-df1f03badb79" (UID: "070d6439-9bd3-4fd8-b61a-df1f03badb79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.225270 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "070d6439-9bd3-4fd8-b61a-df1f03badb79" (UID: "070d6439-9bd3-4fd8-b61a-df1f03badb79"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.229449 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "070d6439-9bd3-4fd8-b61a-df1f03badb79" (UID: "070d6439-9bd3-4fd8-b61a-df1f03badb79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.302839 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.302876 4936 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.302889 4936 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.302900 4936 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/070d6439-9bd3-4fd8-b61a-df1f03badb79-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.302916 4936 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/070d6439-9bd3-4fd8-b61a-df1f03badb79-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.302930 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070d6439-9bd3-4fd8-b61a-df1f03badb79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.302941 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdvw9\" (UniqueName: \"kubernetes.io/projected/070d6439-9bd3-4fd8-b61a-df1f03badb79-kube-api-access-wdvw9\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.824319 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sg9df" event={"ID":"a817ffe2-32a4-4c37-9342-5d642d179ea7","Type":"ContainerStarted","Data":"57e9a22d7c05f84371f34e1628c18bbf7623203688b9e8378a2b8a2195425884"} Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.826604 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6k8ph" event={"ID":"070d6439-9bd3-4fd8-b61a-df1f03badb79","Type":"ContainerDied","Data":"85d0a6cc208da0c19ae0a54a88c4f6c2e0a9f0ec7b253595f38eef9840ae6ae9"} Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.826636 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85d0a6cc208da0c19ae0a54a88c4f6c2e0a9f0ec7b253595f38eef9840ae6ae9" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.826687 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6k8ph" Mar 20 16:18:03 crc kubenswrapper[4936]: I0320 16:18:03.868932 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e353d140-f7e5-47b4-b18e-200d5a255086" path="/var/lib/kubelet/pods/e353d140-f7e5-47b4-b18e-200d5a255086/volumes" Mar 20 16:18:04 crc kubenswrapper[4936]: I0320 16:18:04.233515 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567058-c652h" Mar 20 16:18:04 crc kubenswrapper[4936]: I0320 16:18:04.334423 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vcx7\" (UniqueName: \"kubernetes.io/projected/d274515c-4ea9-459b-b934-ffd2d6982d5c-kube-api-access-8vcx7\") pod \"d274515c-4ea9-459b-b934-ffd2d6982d5c\" (UID: \"d274515c-4ea9-459b-b934-ffd2d6982d5c\") " Mar 20 16:18:04 crc kubenswrapper[4936]: I0320 16:18:04.341777 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d274515c-4ea9-459b-b934-ffd2d6982d5c-kube-api-access-8vcx7" (OuterVolumeSpecName: "kube-api-access-8vcx7") pod "d274515c-4ea9-459b-b934-ffd2d6982d5c" (UID: "d274515c-4ea9-459b-b934-ffd2d6982d5c"). InnerVolumeSpecName "kube-api-access-8vcx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:04 crc kubenswrapper[4936]: I0320 16:18:04.436720 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vcx7\" (UniqueName: \"kubernetes.io/projected/d274515c-4ea9-459b-b934-ffd2d6982d5c-kube-api-access-8vcx7\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:04 crc kubenswrapper[4936]: I0320 16:18:04.866153 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567058-c652h" event={"ID":"d274515c-4ea9-459b-b934-ffd2d6982d5c","Type":"ContainerDied","Data":"f60b06654f5992cfd15834b2fffd483cb501a2737ebd29dfb7ba20d3318ef014"} Mar 20 16:18:04 crc kubenswrapper[4936]: I0320 16:18:04.866501 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f60b06654f5992cfd15834b2fffd483cb501a2737ebd29dfb7ba20d3318ef014" Mar 20 16:18:04 crc kubenswrapper[4936]: I0320 16:18:04.866672 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567058-c652h" Mar 20 16:18:05 crc kubenswrapper[4936]: I0320 16:18:05.300080 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567052-lctck"] Mar 20 16:18:05 crc kubenswrapper[4936]: I0320 16:18:05.306774 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567052-lctck"] Mar 20 16:18:05 crc kubenswrapper[4936]: I0320 16:18:05.867560 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b030e112-1fd3-4be5-a0a7-68ab85e05f93" path="/var/lib/kubelet/pods/b030e112-1fd3-4be5-a0a7-68ab85e05f93/volumes" Mar 20 16:18:06 crc kubenswrapper[4936]: I0320 16:18:06.774456 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rnftv" podUID="cd2037c4-bb05-4f92-9a4a-f71c5da11ebd" containerName="ovn-controller" probeResult="failure" output=< Mar 20 16:18:06 crc kubenswrapper[4936]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 20 16:18:06 crc kubenswrapper[4936]: > Mar 20 16:18:06 crc kubenswrapper[4936]: I0320 16:18:06.831656 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:18:06 crc kubenswrapper[4936]: I0320 16:18:06.841421 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-tp2gk" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.094605 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rnftv-config-zn242"] Mar 20 16:18:07 crc kubenswrapper[4936]: E0320 16:18:07.095261 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070d6439-9bd3-4fd8-b61a-df1f03badb79" containerName="swift-ring-rebalance" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.095373 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="070d6439-9bd3-4fd8-b61a-df1f03badb79" containerName="swift-ring-rebalance" Mar 20 16:18:07 crc kubenswrapper[4936]: E0320 16:18:07.095479 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d274515c-4ea9-459b-b934-ffd2d6982d5c" containerName="oc" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.095583 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="d274515c-4ea9-459b-b934-ffd2d6982d5c" containerName="oc" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.095844 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="d274515c-4ea9-459b-b934-ffd2d6982d5c" containerName="oc" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.095942 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="070d6439-9bd3-4fd8-b61a-df1f03badb79" containerName="swift-ring-rebalance" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.096943 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.099360 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.105192 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rnftv-config-zn242"] Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.199724 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-additional-scripts\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.199864 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.200100 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqb42\" (UniqueName: \"kubernetes.io/projected/7eeb9f43-3253-4a31-8f47-3e40c359e553-kube-api-access-lqb42\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.200222 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-log-ovn\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.200284 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run-ovn\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.200421 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-scripts\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.301756 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run-ovn\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.301834 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-scripts\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.301926 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-additional-scripts\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.301961 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.301999 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqb42\" (UniqueName: \"kubernetes.io/projected/7eeb9f43-3253-4a31-8f47-3e40c359e553-kube-api-access-lqb42\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.302025 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-log-ovn\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.302061 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run-ovn\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.302140 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-log-ovn\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.302199 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.302644 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-additional-scripts\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.304222 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-scripts\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.321353 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqb42\" (UniqueName: \"kubernetes.io/projected/7eeb9f43-3253-4a31-8f47-3e40c359e553-kube-api-access-lqb42\") pod \"ovn-controller-rnftv-config-zn242\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:07 crc kubenswrapper[4936]: I0320 16:18:07.422828 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:08 crc kubenswrapper[4936]: I0320 16:18:08.001859 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rnftv-config-zn242"] Mar 20 16:18:08 crc kubenswrapper[4936]: I0320 16:18:08.906628 4936 generic.go:334] "Generic (PLEG): container finished" podID="7eeb9f43-3253-4a31-8f47-3e40c359e553" containerID="ff195768ebd80a1cba48771ac71507daf8a3ac041103406af17fc990451d3f29" exitCode=0 Mar 20 16:18:08 crc kubenswrapper[4936]: I0320 16:18:08.906709 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rnftv-config-zn242" event={"ID":"7eeb9f43-3253-4a31-8f47-3e40c359e553","Type":"ContainerDied","Data":"ff195768ebd80a1cba48771ac71507daf8a3ac041103406af17fc990451d3f29"} Mar 20 16:18:08 crc kubenswrapper[4936]: I0320 16:18:08.906742 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rnftv-config-zn242" event={"ID":"7eeb9f43-3253-4a31-8f47-3e40c359e553","Type":"ContainerStarted","Data":"f5c55e78619a7eb5457db78c41f9b4df66ecdbc5975c59f6356796be5b4e9dcc"} Mar 20 16:18:11 crc kubenswrapper[4936]: I0320 16:18:11.382781 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 20 16:18:11 crc kubenswrapper[4936]: I0320 16:18:11.802029 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-rnftv" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.191206 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.254950 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-log-ovn\") pod \"7eeb9f43-3253-4a31-8f47-3e40c359e553\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.255031 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run\") pod \"7eeb9f43-3253-4a31-8f47-3e40c359e553\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.255028 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7eeb9f43-3253-4a31-8f47-3e40c359e553" (UID: "7eeb9f43-3253-4a31-8f47-3e40c359e553"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.255095 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7eeb9f43-3253-4a31-8f47-3e40c359e553" (UID: "7eeb9f43-3253-4a31-8f47-3e40c359e553"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.255120 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run" (OuterVolumeSpecName: "var-run") pod "7eeb9f43-3253-4a31-8f47-3e40c359e553" (UID: "7eeb9f43-3253-4a31-8f47-3e40c359e553"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.255057 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run-ovn\") pod \"7eeb9f43-3253-4a31-8f47-3e40c359e553\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.255222 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqb42\" (UniqueName: \"kubernetes.io/projected/7eeb9f43-3253-4a31-8f47-3e40c359e553-kube-api-access-lqb42\") pod \"7eeb9f43-3253-4a31-8f47-3e40c359e553\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.256008 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-additional-scripts\") pod \"7eeb9f43-3253-4a31-8f47-3e40c359e553\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.256150 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-scripts\") pod \"7eeb9f43-3253-4a31-8f47-3e40c359e553\" (UID: \"7eeb9f43-3253-4a31-8f47-3e40c359e553\") " Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.256231 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7eeb9f43-3253-4a31-8f47-3e40c359e553" (UID: "7eeb9f43-3253-4a31-8f47-3e40c359e553"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.256842 4936 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.256865 4936 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.256878 4936 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7eeb9f43-3253-4a31-8f47-3e40c359e553-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.256889 4936 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.257618 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-scripts" (OuterVolumeSpecName: "scripts") pod "7eeb9f43-3253-4a31-8f47-3e40c359e553" (UID: "7eeb9f43-3253-4a31-8f47-3e40c359e553"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.262880 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eeb9f43-3253-4a31-8f47-3e40c359e553-kube-api-access-lqb42" (OuterVolumeSpecName: "kube-api-access-lqb42") pod "7eeb9f43-3253-4a31-8f47-3e40c359e553" (UID: "7eeb9f43-3253-4a31-8f47-3e40c359e553"). InnerVolumeSpecName "kube-api-access-lqb42". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.358823 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eeb9f43-3253-4a31-8f47-3e40c359e553-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.358883 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqb42\" (UniqueName: \"kubernetes.io/projected/7eeb9f43-3253-4a31-8f47-3e40c359e553-kube-api-access-lqb42\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.562831 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.570381 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b1ea233b-41e7-49ac-a878-aa038b350252-etc-swift\") pod \"swift-storage-0\" (UID: \"b1ea233b-41e7-49ac-a878-aa038b350252\") " pod="openstack/swift-storage-0" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.765017 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.974365 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rnftv-config-zn242" event={"ID":"7eeb9f43-3253-4a31-8f47-3e40c359e553","Type":"ContainerDied","Data":"f5c55e78619a7eb5457db78c41f9b4df66ecdbc5975c59f6356796be5b4e9dcc"} Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.974719 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5c55e78619a7eb5457db78c41f9b4df66ecdbc5975c59f6356796be5b4e9dcc" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.974404 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rnftv-config-zn242" Mar 20 16:18:16 crc kubenswrapper[4936]: I0320 16:18:16.986845 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sg9df" event={"ID":"a817ffe2-32a4-4c37-9342-5d642d179ea7","Type":"ContainerStarted","Data":"4e22f1f602d60c236a7c02e8361dd0b77f3c35628b6dc95b2ae098e884f08b6e"} Mar 20 16:18:17 crc kubenswrapper[4936]: I0320 16:18:17.019868 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-sg9df" podStartSLOduration=1.6345341740000001 podStartE2EDuration="15.019845877s" podCreationTimestamp="2026-03-20 16:18:02 +0000 UTC" firstStartedPulling="2026-03-20 16:18:02.995194612 +0000 UTC m=+1033.941562417" lastFinishedPulling="2026-03-20 16:18:16.380506305 +0000 UTC m=+1047.326874120" observedRunningTime="2026-03-20 16:18:17.008479786 +0000 UTC m=+1047.954847621" watchObservedRunningTime="2026-03-20 16:18:17.019845877 +0000 UTC m=+1047.966213692" Mar 20 16:18:17 crc kubenswrapper[4936]: I0320 16:18:17.284273 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rnftv-config-zn242"] Mar 20 16:18:17 crc kubenswrapper[4936]: I0320 16:18:17.294925 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rnftv-config-zn242"] Mar 20 16:18:17 crc kubenswrapper[4936]: W0320 16:18:17.296684 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1ea233b_41e7_49ac_a878_aa038b350252.slice/crio-bd308fea3f84ef93312ad7531d142ea9ff5aa65cc9d42ad84219ade759abd2a9 WatchSource:0}: Error finding container bd308fea3f84ef93312ad7531d142ea9ff5aa65cc9d42ad84219ade759abd2a9: Status 404 returned error can't find the container with id bd308fea3f84ef93312ad7531d142ea9ff5aa65cc9d42ad84219ade759abd2a9 Mar 20 16:18:17 crc kubenswrapper[4936]: I0320 16:18:17.303579 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 20 16:18:17 crc kubenswrapper[4936]: I0320 16:18:17.870004 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eeb9f43-3253-4a31-8f47-3e40c359e553" path="/var/lib/kubelet/pods/7eeb9f43-3253-4a31-8f47-3e40c359e553/volumes" Mar 20 16:18:17 crc kubenswrapper[4936]: I0320 16:18:17.998818 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"bd308fea3f84ef93312ad7531d142ea9ff5aa65cc9d42ad84219ade759abd2a9"} Mar 20 16:18:19 crc kubenswrapper[4936]: I0320 16:18:19.007522 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"c9d4061fb0b0fcd45ed0c535601d8971b93c61f7c474e74f7ca26c408345dc67"} Mar 20 16:18:19 crc kubenswrapper[4936]: I0320 16:18:19.009432 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"dba71ccb8a83c10f2c5daa867170ed990a74602b66ad646469f28bdb266e5f8d"} Mar 20 16:18:20 crc kubenswrapper[4936]: I0320 16:18:20.020843 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"ed4af691ed9f33da15962d2c44dc28e16ff7a8bc4cb12283997ffa4946b59721"} Mar 20 16:18:20 crc kubenswrapper[4936]: I0320 16:18:20.020890 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"122be183c81003684f0116fcf40982367e54ca703488b80e2290c50ae280855a"} Mar 20 16:18:21 crc kubenswrapper[4936]: I0320 16:18:21.034131 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"d6d5cfcdfa801b63a9f3e9454001c0dc8d85bba85d3ed52fc88fd327bf271135"} Mar 20 16:18:21 crc kubenswrapper[4936]: I0320 16:18:21.034776 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"50e55c3fc1076dc740910fac5ae8382f97111d0fa55bc67ccf0e774f2cf9fa82"} Mar 20 16:18:22 crc kubenswrapper[4936]: I0320 16:18:22.045178 4936 generic.go:334] "Generic (PLEG): container finished" podID="18cfd431-66b8-49d8-8dd9-b2184415c006" containerID="3420f7533e5b46917dbbbf04a1c49349ccdcf30014af6ad50b99944a468f9184" exitCode=0 Mar 20 16:18:22 crc kubenswrapper[4936]: I0320 16:18:22.045263 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18cfd431-66b8-49d8-8dd9-b2184415c006","Type":"ContainerDied","Data":"3420f7533e5b46917dbbbf04a1c49349ccdcf30014af6ad50b99944a468f9184"} Mar 20 16:18:22 crc kubenswrapper[4936]: I0320 16:18:22.057664 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"d8bb120bf9212f93e0f329a5952ff07da124c9a27dd0b33b8bda71695ee172e2"} Mar 20 16:18:22 crc kubenswrapper[4936]: I0320 16:18:22.057973 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"ed1a5dfec855a73205444eb04a1bf6284f902382b8131e53841abd56525c3016"} Mar 20 16:18:22 crc kubenswrapper[4936]: I0320 16:18:22.062269 4936 generic.go:334] "Generic (PLEG): container finished" podID="3b3a9af8-80db-4edc-9f6a-42b641c25d71" containerID="5f14a26871add7074b92241d4cf3fd824a81b961fa684c3affe2d0ef82a2d850" exitCode=0 Mar 20 16:18:22 crc kubenswrapper[4936]: I0320 16:18:22.062331 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b3a9af8-80db-4edc-9f6a-42b641c25d71","Type":"ContainerDied","Data":"5f14a26871add7074b92241d4cf3fd824a81b961fa684c3affe2d0ef82a2d850"} Mar 20 16:18:25 crc kubenswrapper[4936]: I0320 16:18:25.098508 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18cfd431-66b8-49d8-8dd9-b2184415c006","Type":"ContainerStarted","Data":"5339d13649d67f88dad82aff5c8edfb374e4b7801e2e05f61ef75ccb072a1b84"} Mar 20 16:18:25 crc kubenswrapper[4936]: I0320 16:18:25.099534 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:18:25 crc kubenswrapper[4936]: I0320 16:18:25.102906 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b3a9af8-80db-4edc-9f6a-42b641c25d71","Type":"ContainerStarted","Data":"a060a93f3074c738cf46afdebfc8990015690cbb054bb0cf67be5374bce28bde"} Mar 20 16:18:25 crc kubenswrapper[4936]: I0320 16:18:25.103641 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 20 16:18:25 crc kubenswrapper[4936]: I0320 16:18:25.123303 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.141211174 podStartE2EDuration="1m29.123285659s" podCreationTimestamp="2026-03-20 16:16:56 +0000 UTC" firstStartedPulling="2026-03-20 16:16:58.681235957 +0000 UTC m=+969.627603762" lastFinishedPulling="2026-03-20 16:17:48.663310432 +0000 UTC m=+1019.609678247" observedRunningTime="2026-03-20 16:18:25.122269669 +0000 UTC m=+1056.068637484" watchObservedRunningTime="2026-03-20 16:18:25.123285659 +0000 UTC m=+1056.069653474" Mar 20 16:18:25 crc kubenswrapper[4936]: I0320 16:18:25.167381 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.590813931 podStartE2EDuration="1m29.167358763s" podCreationTimestamp="2026-03-20 16:16:56 +0000 UTC" firstStartedPulling="2026-03-20 16:16:58.048585897 +0000 UTC m=+968.994953712" lastFinishedPulling="2026-03-20 16:17:48.625130739 +0000 UTC m=+1019.571498544" observedRunningTime="2026-03-20 16:18:25.160951796 +0000 UTC m=+1056.107319621" watchObservedRunningTime="2026-03-20 16:18:25.167358763 +0000 UTC m=+1056.113726578" Mar 20 16:18:25 crc kubenswrapper[4936]: I0320 16:18:25.863692 4936 scope.go:117] "RemoveContainer" containerID="836dcf855452eef8c5987de7d0547f19402e63dbed8f386726c2544337c003b7" Mar 20 16:18:28 crc kubenswrapper[4936]: I0320 16:18:28.136129 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"e53c05ca31e150f662a6a26a0584162b0beaddec1c3c366aae95e5945b0dc174"} Mar 20 16:18:28 crc kubenswrapper[4936]: I0320 16:18:28.136532 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"03fa9e459acf6d3e594da6811eafec85111b0ac7eb1f9a616dca68e96b045d28"} Mar 20 16:18:28 crc kubenswrapper[4936]: I0320 16:18:28.136558 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"79cebc85f5ca3303a290687115533f5e757315409c8a8c5c130d1a2ae6c94727"} Mar 20 16:18:28 crc kubenswrapper[4936]: I0320 16:18:28.136567 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"a6a878bcc274f05a11dd7934008d6f943ae3ff412755fc626da729bd68e94d03"} Mar 20 16:18:28 crc kubenswrapper[4936]: I0320 16:18:28.136576 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"8cf7b046279956bc8c5730b8f36c78818d1f4f1bd0ac20fbccaaea5416ca5f64"} Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.164342 4936 generic.go:334] "Generic (PLEG): container finished" podID="a817ffe2-32a4-4c37-9342-5d642d179ea7" containerID="4e22f1f602d60c236a7c02e8361dd0b77f3c35628b6dc95b2ae098e884f08b6e" exitCode=0 Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.164504 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sg9df" event={"ID":"a817ffe2-32a4-4c37-9342-5d642d179ea7","Type":"ContainerDied","Data":"4e22f1f602d60c236a7c02e8361dd0b77f3c35628b6dc95b2ae098e884f08b6e"} Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.174602 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"6740a91f06041cde57cefcd8cd3a6c90017d240676814edd12332cbf9e0a582c"} Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.174655 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b1ea233b-41e7-49ac-a878-aa038b350252","Type":"ContainerStarted","Data":"25d05e1cab6703b0523958bb053a9f6fa6857dd5661e3df54189262bbe4ad9a8"} Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.228368 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.276958594999996 podStartE2EDuration="46.228351289s" podCreationTimestamp="2026-03-20 16:17:43 +0000 UTC" firstStartedPulling="2026-03-20 16:18:17.29891627 +0000 UTC m=+1048.245284085" lastFinishedPulling="2026-03-20 16:18:27.250308964 +0000 UTC m=+1058.196676779" observedRunningTime="2026-03-20 16:18:29.225375342 +0000 UTC m=+1060.171743167" watchObservedRunningTime="2026-03-20 16:18:29.228351289 +0000 UTC m=+1060.174719104" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.555285 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-szfvw"] Mar 20 16:18:29 crc kubenswrapper[4936]: E0320 16:18:29.555620 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eeb9f43-3253-4a31-8f47-3e40c359e553" containerName="ovn-config" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.555633 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eeb9f43-3253-4a31-8f47-3e40c359e553" containerName="ovn-config" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.555801 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eeb9f43-3253-4a31-8f47-3e40c359e553" containerName="ovn-config" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.556560 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.560846 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.573519 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-szfvw"] Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.581721 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.581778 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-config\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.581815 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.581854 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.582190 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnzsq\" (UniqueName: \"kubernetes.io/projected/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-kube-api-access-lnzsq\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.582278 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-svc\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.684236 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnzsq\" (UniqueName: \"kubernetes.io/projected/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-kube-api-access-lnzsq\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.684333 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-svc\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.684399 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.684423 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-config\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.684456 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.684488 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.685585 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.685613 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.686178 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-config\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.686271 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-svc\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.686796 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.710439 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnzsq\" (UniqueName: \"kubernetes.io/projected/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-kube-api-access-lnzsq\") pod \"dnsmasq-dns-764c5664d7-szfvw\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:29 crc kubenswrapper[4936]: I0320 16:18:29.876730 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.317093 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-szfvw"] Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.550459 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.699013 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-db-sync-config-data\") pod \"a817ffe2-32a4-4c37-9342-5d642d179ea7\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.699091 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drsfw\" (UniqueName: \"kubernetes.io/projected/a817ffe2-32a4-4c37-9342-5d642d179ea7-kube-api-access-drsfw\") pod \"a817ffe2-32a4-4c37-9342-5d642d179ea7\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.699300 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-combined-ca-bundle\") pod \"a817ffe2-32a4-4c37-9342-5d642d179ea7\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.699357 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-config-data\") pod \"a817ffe2-32a4-4c37-9342-5d642d179ea7\" (UID: \"a817ffe2-32a4-4c37-9342-5d642d179ea7\") " Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.703308 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a817ffe2-32a4-4c37-9342-5d642d179ea7" (UID: "a817ffe2-32a4-4c37-9342-5d642d179ea7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.703498 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a817ffe2-32a4-4c37-9342-5d642d179ea7-kube-api-access-drsfw" (OuterVolumeSpecName: "kube-api-access-drsfw") pod "a817ffe2-32a4-4c37-9342-5d642d179ea7" (UID: "a817ffe2-32a4-4c37-9342-5d642d179ea7"). InnerVolumeSpecName "kube-api-access-drsfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.721411 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a817ffe2-32a4-4c37-9342-5d642d179ea7" (UID: "a817ffe2-32a4-4c37-9342-5d642d179ea7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.747950 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-config-data" (OuterVolumeSpecName: "config-data") pod "a817ffe2-32a4-4c37-9342-5d642d179ea7" (UID: "a817ffe2-32a4-4c37-9342-5d642d179ea7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.801432 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.801466 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.801476 4936 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a817ffe2-32a4-4c37-9342-5d642d179ea7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:30 crc kubenswrapper[4936]: I0320 16:18:30.801487 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drsfw\" (UniqueName: \"kubernetes.io/projected/a817ffe2-32a4-4c37-9342-5d642d179ea7-kube-api-access-drsfw\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.191590 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sg9df" event={"ID":"a817ffe2-32a4-4c37-9342-5d642d179ea7","Type":"ContainerDied","Data":"57e9a22d7c05f84371f34e1628c18bbf7623203688b9e8378a2b8a2195425884"} Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.191948 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57e9a22d7c05f84371f34e1628c18bbf7623203688b9e8378a2b8a2195425884" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.191648 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sg9df" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.194009 4936 generic.go:334] "Generic (PLEG): container finished" podID="9e7a727d-a8e0-49b3-8246-614fbf56e1dd" containerID="ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452" exitCode=0 Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.194042 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" event={"ID":"9e7a727d-a8e0-49b3-8246-614fbf56e1dd","Type":"ContainerDied","Data":"ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452"} Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.194066 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" event={"ID":"9e7a727d-a8e0-49b3-8246-614fbf56e1dd","Type":"ContainerStarted","Data":"81f38c152ce3aaf601e857c2f25b2811d1b902eaacd515f355133e1b88f1fa5a"} Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.606494 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-szfvw"] Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.641441 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-6g4tb"] Mar 20 16:18:31 crc kubenswrapper[4936]: E0320 16:18:31.641870 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a817ffe2-32a4-4c37-9342-5d642d179ea7" containerName="glance-db-sync" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.641896 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a817ffe2-32a4-4c37-9342-5d642d179ea7" containerName="glance-db-sync" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.642161 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a817ffe2-32a4-4c37-9342-5d642d179ea7" containerName="glance-db-sync" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.643224 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.653860 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-6g4tb"] Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.818350 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-476wm\" (UniqueName: \"kubernetes.io/projected/f53878d4-2d6f-45c0-a7d8-a143144d5061-kube-api-access-476wm\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.818694 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.818757 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.818793 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-config\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.818813 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.818917 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.920882 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-config\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.920933 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.921016 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.921073 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-476wm\" (UniqueName: \"kubernetes.io/projected/f53878d4-2d6f-45c0-a7d8-a143144d5061-kube-api-access-476wm\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.921097 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.921137 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.921892 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-config\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.921917 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.922394 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.922459 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.922498 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.948190 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-476wm\" (UniqueName: \"kubernetes.io/projected/f53878d4-2d6f-45c0-a7d8-a143144d5061-kube-api-access-476wm\") pod \"dnsmasq-dns-74f6bcbc87-6g4tb\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:31 crc kubenswrapper[4936]: I0320 16:18:31.967930 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:32 crc kubenswrapper[4936]: I0320 16:18:32.242595 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" event={"ID":"9e7a727d-a8e0-49b3-8246-614fbf56e1dd","Type":"ContainerStarted","Data":"6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594"} Mar 20 16:18:32 crc kubenswrapper[4936]: I0320 16:18:32.242915 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:32 crc kubenswrapper[4936]: I0320 16:18:32.300662 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" podStartSLOduration=3.300641451 podStartE2EDuration="3.300641451s" podCreationTimestamp="2026-03-20 16:18:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:18:32.300140127 +0000 UTC m=+1063.246507942" watchObservedRunningTime="2026-03-20 16:18:32.300641451 +0000 UTC m=+1063.247009266" Mar 20 16:18:32 crc kubenswrapper[4936]: I0320 16:18:32.442140 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-6g4tb"] Mar 20 16:18:32 crc kubenswrapper[4936]: W0320 16:18:32.446459 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf53878d4_2d6f_45c0_a7d8_a143144d5061.slice/crio-8ff681eddc0a3906d98c4446488b70a2cdec3e5968cb43dd86e6c3bc3547768e WatchSource:0}: Error finding container 8ff681eddc0a3906d98c4446488b70a2cdec3e5968cb43dd86e6c3bc3547768e: Status 404 returned error can't find the container with id 8ff681eddc0a3906d98c4446488b70a2cdec3e5968cb43dd86e6c3bc3547768e Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.252111 4936 generic.go:334] "Generic (PLEG): container finished" podID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerID="2d8acde0827e044d41df774e7e14759e88630ce148f1aca6d31bcfbc0fe8c3c9" exitCode=0 Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.252173 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" event={"ID":"f53878d4-2d6f-45c0-a7d8-a143144d5061","Type":"ContainerDied","Data":"2d8acde0827e044d41df774e7e14759e88630ce148f1aca6d31bcfbc0fe8c3c9"} Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.252503 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" event={"ID":"f53878d4-2d6f-45c0-a7d8-a143144d5061","Type":"ContainerStarted","Data":"8ff681eddc0a3906d98c4446488b70a2cdec3e5968cb43dd86e6c3bc3547768e"} Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.252926 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" podUID="9e7a727d-a8e0-49b3-8246-614fbf56e1dd" containerName="dnsmasq-dns" containerID="cri-o://6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594" gracePeriod=10 Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.645264 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.753255 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnzsq\" (UniqueName: \"kubernetes.io/projected/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-kube-api-access-lnzsq\") pod \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.753296 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-svc\") pod \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.753356 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-sb\") pod \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.753454 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-swift-storage-0\") pod \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.753485 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-config\") pod \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.753535 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-nb\") pod \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\" (UID: \"9e7a727d-a8e0-49b3-8246-614fbf56e1dd\") " Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.760874 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-kube-api-access-lnzsq" (OuterVolumeSpecName: "kube-api-access-lnzsq") pod "9e7a727d-a8e0-49b3-8246-614fbf56e1dd" (UID: "9e7a727d-a8e0-49b3-8246-614fbf56e1dd"). InnerVolumeSpecName "kube-api-access-lnzsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.809613 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9e7a727d-a8e0-49b3-8246-614fbf56e1dd" (UID: "9e7a727d-a8e0-49b3-8246-614fbf56e1dd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.814940 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-config" (OuterVolumeSpecName: "config") pod "9e7a727d-a8e0-49b3-8246-614fbf56e1dd" (UID: "9e7a727d-a8e0-49b3-8246-614fbf56e1dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.816113 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9e7a727d-a8e0-49b3-8246-614fbf56e1dd" (UID: "9e7a727d-a8e0-49b3-8246-614fbf56e1dd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.816531 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e7a727d-a8e0-49b3-8246-614fbf56e1dd" (UID: "9e7a727d-a8e0-49b3-8246-614fbf56e1dd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.824451 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e7a727d-a8e0-49b3-8246-614fbf56e1dd" (UID: "9e7a727d-a8e0-49b3-8246-614fbf56e1dd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.855457 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.855493 4936 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.855503 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.855511 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.855520 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnzsq\" (UniqueName: \"kubernetes.io/projected/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-kube-api-access-lnzsq\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:33 crc kubenswrapper[4936]: I0320 16:18:33.855739 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e7a727d-a8e0-49b3-8246-614fbf56e1dd-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.261766 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" event={"ID":"f53878d4-2d6f-45c0-a7d8-a143144d5061","Type":"ContainerStarted","Data":"8e03a6e951063b53256be6fad58cf93a73368d0fa4111e4c8bff3041a956b205"} Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.261917 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.264161 4936 generic.go:334] "Generic (PLEG): container finished" podID="9e7a727d-a8e0-49b3-8246-614fbf56e1dd" containerID="6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594" exitCode=0 Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.264211 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" event={"ID":"9e7a727d-a8e0-49b3-8246-614fbf56e1dd","Type":"ContainerDied","Data":"6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594"} Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.264240 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" event={"ID":"9e7a727d-a8e0-49b3-8246-614fbf56e1dd","Type":"ContainerDied","Data":"81f38c152ce3aaf601e857c2f25b2811d1b902eaacd515f355133e1b88f1fa5a"} Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.264258 4936 scope.go:117] "RemoveContainer" containerID="6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594" Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.264394 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-szfvw" Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.283423 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" podStartSLOduration=3.283403033 podStartE2EDuration="3.283403033s" podCreationTimestamp="2026-03-20 16:18:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:18:34.281021033 +0000 UTC m=+1065.227388888" watchObservedRunningTime="2026-03-20 16:18:34.283403033 +0000 UTC m=+1065.229770848" Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.287062 4936 scope.go:117] "RemoveContainer" containerID="ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452" Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.302041 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-szfvw"] Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.309148 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-szfvw"] Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.312408 4936 scope.go:117] "RemoveContainer" containerID="6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594" Mar 20 16:18:34 crc kubenswrapper[4936]: E0320 16:18:34.312901 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594\": container with ID starting with 6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594 not found: ID does not exist" containerID="6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594" Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.312941 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594"} err="failed to get container status \"6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594\": rpc error: code = NotFound desc = could not find container \"6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594\": container with ID starting with 6aa177e7cc1e207a1ae5231aa6802f3bf75cefa400971c86ab70e84fef6d3594 not found: ID does not exist" Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.312970 4936 scope.go:117] "RemoveContainer" containerID="ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452" Mar 20 16:18:34 crc kubenswrapper[4936]: E0320 16:18:34.313470 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452\": container with ID starting with ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452 not found: ID does not exist" containerID="ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452" Mar 20 16:18:34 crc kubenswrapper[4936]: I0320 16:18:34.313509 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452"} err="failed to get container status \"ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452\": rpc error: code = NotFound desc = could not find container \"ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452\": container with ID starting with ca6899c2a0036e6b4fdb0a3fb5c71dc8dacc678148789c147cf58bf91b04f452 not found: ID does not exist" Mar 20 16:18:35 crc kubenswrapper[4936]: I0320 16:18:35.864305 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e7a727d-a8e0-49b3-8246-614fbf56e1dd" path="/var/lib/kubelet/pods/9e7a727d-a8e0-49b3-8246-614fbf56e1dd/volumes" Mar 20 16:18:37 crc kubenswrapper[4936]: I0320 16:18:37.562821 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="3b3a9af8-80db-4edc-9f6a-42b641c25d71" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Mar 20 16:18:37 crc kubenswrapper[4936]: I0320 16:18:37.974122 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="18cfd431-66b8-49d8-8dd9-b2184415c006" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Mar 20 16:18:41 crc kubenswrapper[4936]: I0320 16:18:41.970850 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.031100 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-22blq"] Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.031454 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-22blq" podUID="223bcfee-4f1a-4ce5-9e38-a8bd4676411b" containerName="dnsmasq-dns" containerID="cri-o://7cec49d48d8b5e1fcfe1151b1cb1485df8291075bb1345b9aa4d79fa5e372060" gracePeriod=10 Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.346637 4936 generic.go:334] "Generic (PLEG): container finished" podID="223bcfee-4f1a-4ce5-9e38-a8bd4676411b" containerID="7cec49d48d8b5e1fcfe1151b1cb1485df8291075bb1345b9aa4d79fa5e372060" exitCode=0 Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.346864 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-22blq" event={"ID":"223bcfee-4f1a-4ce5-9e38-a8bd4676411b","Type":"ContainerDied","Data":"7cec49d48d8b5e1fcfe1151b1cb1485df8291075bb1345b9aa4d79fa5e372060"} Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.596369 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.728515 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-config\") pod \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.728958 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzdsk\" (UniqueName: \"kubernetes.io/projected/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-kube-api-access-wzdsk\") pod \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.729095 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-nb\") pod \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.729230 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-sb\") pod \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.729337 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-dns-svc\") pod \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\" (UID: \"223bcfee-4f1a-4ce5-9e38-a8bd4676411b\") " Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.741422 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-kube-api-access-wzdsk" (OuterVolumeSpecName: "kube-api-access-wzdsk") pod "223bcfee-4f1a-4ce5-9e38-a8bd4676411b" (UID: "223bcfee-4f1a-4ce5-9e38-a8bd4676411b"). InnerVolumeSpecName "kube-api-access-wzdsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.766377 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-config" (OuterVolumeSpecName: "config") pod "223bcfee-4f1a-4ce5-9e38-a8bd4676411b" (UID: "223bcfee-4f1a-4ce5-9e38-a8bd4676411b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.766401 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "223bcfee-4f1a-4ce5-9e38-a8bd4676411b" (UID: "223bcfee-4f1a-4ce5-9e38-a8bd4676411b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.766427 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "223bcfee-4f1a-4ce5-9e38-a8bd4676411b" (UID: "223bcfee-4f1a-4ce5-9e38-a8bd4676411b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.769067 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "223bcfee-4f1a-4ce5-9e38-a8bd4676411b" (UID: "223bcfee-4f1a-4ce5-9e38-a8bd4676411b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.831716 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.831754 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzdsk\" (UniqueName: \"kubernetes.io/projected/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-kube-api-access-wzdsk\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.831766 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.831775 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:42 crc kubenswrapper[4936]: I0320 16:18:42.831783 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/223bcfee-4f1a-4ce5-9e38-a8bd4676411b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:43 crc kubenswrapper[4936]: I0320 16:18:43.355713 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-22blq" Mar 20 16:18:43 crc kubenswrapper[4936]: I0320 16:18:43.355778 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-22blq" event={"ID":"223bcfee-4f1a-4ce5-9e38-a8bd4676411b","Type":"ContainerDied","Data":"162fc1cbc8c40839a51a85bd28799da36f43202906d9030242282e0ca56a8d9f"} Mar 20 16:18:43 crc kubenswrapper[4936]: I0320 16:18:43.357734 4936 scope.go:117] "RemoveContainer" containerID="7cec49d48d8b5e1fcfe1151b1cb1485df8291075bb1345b9aa4d79fa5e372060" Mar 20 16:18:43 crc kubenswrapper[4936]: I0320 16:18:43.379777 4936 scope.go:117] "RemoveContainer" containerID="59a9b9149a96aa019822cbac94464d9c9c9b540e526b2838ca44aac6399da5fd" Mar 20 16:18:43 crc kubenswrapper[4936]: I0320 16:18:43.396231 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-22blq"] Mar 20 16:18:43 crc kubenswrapper[4936]: I0320 16:18:43.404905 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-22blq"] Mar 20 16:18:43 crc kubenswrapper[4936]: I0320 16:18:43.870747 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="223bcfee-4f1a-4ce5-9e38-a8bd4676411b" path="/var/lib/kubelet/pods/223bcfee-4f1a-4ce5-9e38-a8bd4676411b/volumes" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.561253 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.921376 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-p249h"] Mar 20 16:18:47 crc kubenswrapper[4936]: E0320 16:18:47.921780 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e7a727d-a8e0-49b3-8246-614fbf56e1dd" containerName="init" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.921805 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e7a727d-a8e0-49b3-8246-614fbf56e1dd" containerName="init" Mar 20 16:18:47 crc kubenswrapper[4936]: E0320 16:18:47.921839 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223bcfee-4f1a-4ce5-9e38-a8bd4676411b" containerName="init" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.921847 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="223bcfee-4f1a-4ce5-9e38-a8bd4676411b" containerName="init" Mar 20 16:18:47 crc kubenswrapper[4936]: E0320 16:18:47.921864 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223bcfee-4f1a-4ce5-9e38-a8bd4676411b" containerName="dnsmasq-dns" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.921871 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="223bcfee-4f1a-4ce5-9e38-a8bd4676411b" containerName="dnsmasq-dns" Mar 20 16:18:47 crc kubenswrapper[4936]: E0320 16:18:47.921879 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e7a727d-a8e0-49b3-8246-614fbf56e1dd" containerName="dnsmasq-dns" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.921886 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e7a727d-a8e0-49b3-8246-614fbf56e1dd" containerName="dnsmasq-dns" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.922082 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e7a727d-a8e0-49b3-8246-614fbf56e1dd" containerName="dnsmasq-dns" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.922100 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="223bcfee-4f1a-4ce5-9e38-a8bd4676411b" containerName="dnsmasq-dns" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.922715 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-p249h" Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.932251 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-p249h"] Mar 20 16:18:47 crc kubenswrapper[4936]: I0320 16:18:47.967728 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.018196 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4dh8\" (UniqueName: \"kubernetes.io/projected/2726b8ff-699c-419e-ae52-0868c355ea23-kube-api-access-l4dh8\") pod \"cinder-db-create-p249h\" (UID: \"2726b8ff-699c-419e-ae52-0868c355ea23\") " pod="openstack/cinder-db-create-p249h" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.018316 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2726b8ff-699c-419e-ae52-0868c355ea23-operator-scripts\") pod \"cinder-db-create-p249h\" (UID: \"2726b8ff-699c-419e-ae52-0868c355ea23\") " pod="openstack/cinder-db-create-p249h" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.053830 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-63a4-account-create-update-wng8g"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.054963 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.059045 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.062258 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-63a4-account-create-update-wng8g"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.121480 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2726b8ff-699c-419e-ae52-0868c355ea23-operator-scripts\") pod \"cinder-db-create-p249h\" (UID: \"2726b8ff-699c-419e-ae52-0868c355ea23\") " pod="openstack/cinder-db-create-p249h" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.121658 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4dh8\" (UniqueName: \"kubernetes.io/projected/2726b8ff-699c-419e-ae52-0868c355ea23-kube-api-access-l4dh8\") pod \"cinder-db-create-p249h\" (UID: \"2726b8ff-699c-419e-ae52-0868c355ea23\") " pod="openstack/cinder-db-create-p249h" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.123358 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2726b8ff-699c-419e-ae52-0868c355ea23-operator-scripts\") pod \"cinder-db-create-p249h\" (UID: \"2726b8ff-699c-419e-ae52-0868c355ea23\") " pod="openstack/cinder-db-create-p249h" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.128587 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-s46fs"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.131078 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.140226 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-s46fs"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.155496 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4dh8\" (UniqueName: \"kubernetes.io/projected/2726b8ff-699c-419e-ae52-0868c355ea23-kube-api-access-l4dh8\") pod \"cinder-db-create-p249h\" (UID: \"2726b8ff-699c-419e-ae52-0868c355ea23\") " pod="openstack/cinder-db-create-p249h" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.205700 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wfxxc"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.206924 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.211430 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gbshp" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.211463 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.211648 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.211861 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.224767 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0288b360-6ff9-4568-bc31-1351a3e55ad2-operator-scripts\") pod \"barbican-db-create-s46fs\" (UID: \"0288b360-6ff9-4568-bc31-1351a3e55ad2\") " pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.224916 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df8f8d6e-b8f0-4148-940e-712399f97ab1-operator-scripts\") pod \"cinder-63a4-account-create-update-wng8g\" (UID: \"df8f8d6e-b8f0-4148-940e-712399f97ab1\") " pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.224945 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s85s\" (UniqueName: \"kubernetes.io/projected/df8f8d6e-b8f0-4148-940e-712399f97ab1-kube-api-access-2s85s\") pod \"cinder-63a4-account-create-update-wng8g\" (UID: \"df8f8d6e-b8f0-4148-940e-712399f97ab1\") " pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.224978 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kx6m\" (UniqueName: \"kubernetes.io/projected/0288b360-6ff9-4568-bc31-1351a3e55ad2-kube-api-access-8kx6m\") pod \"barbican-db-create-s46fs\" (UID: \"0288b360-6ff9-4568-bc31-1351a3e55ad2\") " pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.239083 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wfxxc"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.243236 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-p249h" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.252601 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ab53-account-create-update-rrk7z"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.254148 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.258310 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.289358 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ab53-account-create-update-rrk7z"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.326579 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b802c0e2-15be-4604-b64c-3b7441a1096d-operator-scripts\") pod \"neutron-ab53-account-create-update-rrk7z\" (UID: \"b802c0e2-15be-4604-b64c-3b7441a1096d\") " pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.326946 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0288b360-6ff9-4568-bc31-1351a3e55ad2-operator-scripts\") pod \"barbican-db-create-s46fs\" (UID: \"0288b360-6ff9-4568-bc31-1351a3e55ad2\") " pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.327071 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbp4c\" (UniqueName: \"kubernetes.io/projected/b802c0e2-15be-4604-b64c-3b7441a1096d-kube-api-access-jbp4c\") pod \"neutron-ab53-account-create-update-rrk7z\" (UID: \"b802c0e2-15be-4604-b64c-3b7441a1096d\") " pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.327376 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-config-data\") pod \"keystone-db-sync-wfxxc\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.327486 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-combined-ca-bundle\") pod \"keystone-db-sync-wfxxc\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.327656 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb7qk\" (UniqueName: \"kubernetes.io/projected/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-kube-api-access-cb7qk\") pod \"keystone-db-sync-wfxxc\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.327866 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df8f8d6e-b8f0-4148-940e-712399f97ab1-operator-scripts\") pod \"cinder-63a4-account-create-update-wng8g\" (UID: \"df8f8d6e-b8f0-4148-940e-712399f97ab1\") " pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.327992 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s85s\" (UniqueName: \"kubernetes.io/projected/df8f8d6e-b8f0-4148-940e-712399f97ab1-kube-api-access-2s85s\") pod \"cinder-63a4-account-create-update-wng8g\" (UID: \"df8f8d6e-b8f0-4148-940e-712399f97ab1\") " pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.328130 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0288b360-6ff9-4568-bc31-1351a3e55ad2-operator-scripts\") pod \"barbican-db-create-s46fs\" (UID: \"0288b360-6ff9-4568-bc31-1351a3e55ad2\") " pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.328237 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kx6m\" (UniqueName: \"kubernetes.io/projected/0288b360-6ff9-4568-bc31-1351a3e55ad2-kube-api-access-8kx6m\") pod \"barbican-db-create-s46fs\" (UID: \"0288b360-6ff9-4568-bc31-1351a3e55ad2\") " pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.328741 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df8f8d6e-b8f0-4148-940e-712399f97ab1-operator-scripts\") pod \"cinder-63a4-account-create-update-wng8g\" (UID: \"df8f8d6e-b8f0-4148-940e-712399f97ab1\") " pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.329491 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-m5pnp"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.330566 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.338236 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-m5pnp"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.344280 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7437-account-create-update-qn4hp"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.345329 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.347193 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.362559 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kx6m\" (UniqueName: \"kubernetes.io/projected/0288b360-6ff9-4568-bc31-1351a3e55ad2-kube-api-access-8kx6m\") pod \"barbican-db-create-s46fs\" (UID: \"0288b360-6ff9-4568-bc31-1351a3e55ad2\") " pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.369948 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s85s\" (UniqueName: \"kubernetes.io/projected/df8f8d6e-b8f0-4148-940e-712399f97ab1-kube-api-access-2s85s\") pod \"cinder-63a4-account-create-update-wng8g\" (UID: \"df8f8d6e-b8f0-4148-940e-712399f97ab1\") " pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.370589 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7437-account-create-update-qn4hp"] Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.377993 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.430371 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbp4c\" (UniqueName: \"kubernetes.io/projected/b802c0e2-15be-4604-b64c-3b7441a1096d-kube-api-access-jbp4c\") pod \"neutron-ab53-account-create-update-rrk7z\" (UID: \"b802c0e2-15be-4604-b64c-3b7441a1096d\") " pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.430616 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-config-data\") pod \"keystone-db-sync-wfxxc\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.430636 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-combined-ca-bundle\") pod \"keystone-db-sync-wfxxc\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.430654 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gtnx\" (UniqueName: \"kubernetes.io/projected/da80eafb-7c87-4b87-89ae-beee42d9b256-kube-api-access-8gtnx\") pod \"neutron-db-create-m5pnp\" (UID: \"da80eafb-7c87-4b87-89ae-beee42d9b256\") " pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.430694 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb7qk\" (UniqueName: \"kubernetes.io/projected/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-kube-api-access-cb7qk\") pod \"keystone-db-sync-wfxxc\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.430713 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-operator-scripts\") pod \"barbican-7437-account-create-update-qn4hp\" (UID: \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\") " pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.430792 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da80eafb-7c87-4b87-89ae-beee42d9b256-operator-scripts\") pod \"neutron-db-create-m5pnp\" (UID: \"da80eafb-7c87-4b87-89ae-beee42d9b256\") " pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.430846 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwcwq\" (UniqueName: \"kubernetes.io/projected/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-kube-api-access-nwcwq\") pod \"barbican-7437-account-create-update-qn4hp\" (UID: \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\") " pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.430877 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b802c0e2-15be-4604-b64c-3b7441a1096d-operator-scripts\") pod \"neutron-ab53-account-create-update-rrk7z\" (UID: \"b802c0e2-15be-4604-b64c-3b7441a1096d\") " pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.431693 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b802c0e2-15be-4604-b64c-3b7441a1096d-operator-scripts\") pod \"neutron-ab53-account-create-update-rrk7z\" (UID: \"b802c0e2-15be-4604-b64c-3b7441a1096d\") " pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.435419 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-config-data\") pod \"keystone-db-sync-wfxxc\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.438918 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-combined-ca-bundle\") pod \"keystone-db-sync-wfxxc\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.448146 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbp4c\" (UniqueName: \"kubernetes.io/projected/b802c0e2-15be-4604-b64c-3b7441a1096d-kube-api-access-jbp4c\") pod \"neutron-ab53-account-create-update-rrk7z\" (UID: \"b802c0e2-15be-4604-b64c-3b7441a1096d\") " pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.449375 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb7qk\" (UniqueName: \"kubernetes.io/projected/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-kube-api-access-cb7qk\") pod \"keystone-db-sync-wfxxc\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.502438 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:48 crc kubenswrapper[4936]: I0320 16:18:48.541243 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.543143 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwcwq\" (UniqueName: \"kubernetes.io/projected/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-kube-api-access-nwcwq\") pod \"barbican-7437-account-create-update-qn4hp\" (UID: \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\") " pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.543281 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gtnx\" (UniqueName: \"kubernetes.io/projected/da80eafb-7c87-4b87-89ae-beee42d9b256-kube-api-access-8gtnx\") pod \"neutron-db-create-m5pnp\" (UID: \"da80eafb-7c87-4b87-89ae-beee42d9b256\") " pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.543371 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-operator-scripts\") pod \"barbican-7437-account-create-update-qn4hp\" (UID: \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\") " pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.543679 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da80eafb-7c87-4b87-89ae-beee42d9b256-operator-scripts\") pod \"neutron-db-create-m5pnp\" (UID: \"da80eafb-7c87-4b87-89ae-beee42d9b256\") " pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.544562 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-operator-scripts\") pod \"barbican-7437-account-create-update-qn4hp\" (UID: \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\") " pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.544574 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da80eafb-7c87-4b87-89ae-beee42d9b256-operator-scripts\") pod \"neutron-db-create-m5pnp\" (UID: \"da80eafb-7c87-4b87-89ae-beee42d9b256\") " pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.560455 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gtnx\" (UniqueName: \"kubernetes.io/projected/da80eafb-7c87-4b87-89ae-beee42d9b256-kube-api-access-8gtnx\") pod \"neutron-db-create-m5pnp\" (UID: \"da80eafb-7c87-4b87-89ae-beee42d9b256\") " pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.560646 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwcwq\" (UniqueName: \"kubernetes.io/projected/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-kube-api-access-nwcwq\") pod \"barbican-7437-account-create-update-qn4hp\" (UID: \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\") " pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.653865 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.762953 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.775734 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:48.816904 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-p249h"] Mar 20 16:18:49 crc kubenswrapper[4936]: W0320 16:18:48.825969 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2726b8ff_699c_419e_ae52_0868c355ea23.slice/crio-c5035aec792808e37a45d18ecf0672ad339bafa9b403c12682a47dd0d3179748 WatchSource:0}: Error finding container c5035aec792808e37a45d18ecf0672ad339bafa9b403c12682a47dd0d3179748: Status 404 returned error can't find the container with id c5035aec792808e37a45d18ecf0672ad339bafa9b403c12682a47dd0d3179748 Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.415375 4936 generic.go:334] "Generic (PLEG): container finished" podID="2726b8ff-699c-419e-ae52-0868c355ea23" containerID="6708d423b696ff66673a16646000716783814ae4031db65a5f25877b13a8bae9" exitCode=0 Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.415738 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-p249h" event={"ID":"2726b8ff-699c-419e-ae52-0868c355ea23","Type":"ContainerDied","Data":"6708d423b696ff66673a16646000716783814ae4031db65a5f25877b13a8bae9"} Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.415939 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-p249h" event={"ID":"2726b8ff-699c-419e-ae52-0868c355ea23","Type":"ContainerStarted","Data":"c5035aec792808e37a45d18ecf0672ad339bafa9b403c12682a47dd0d3179748"} Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.506247 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-63a4-account-create-update-wng8g"] Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.588380 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-s46fs"] Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.600910 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wfxxc"] Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.610977 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ab53-account-create-update-rrk7z"] Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.617367 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7437-account-create-update-qn4hp"] Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.623493 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-m5pnp"] Mar 20 16:18:49 crc kubenswrapper[4936]: I0320 16:18:49.628715 4936 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 16:18:49 crc kubenswrapper[4936]: W0320 16:18:49.639727 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0288b360_6ff9_4568_bc31_1351a3e55ad2.slice/crio-bc1638d382f5d8093cc871d156cf7294fee1f80b81c87c9ebbb1dff49490b230 WatchSource:0}: Error finding container bc1638d382f5d8093cc871d156cf7294fee1f80b81c87c9ebbb1dff49490b230: Status 404 returned error can't find the container with id bc1638d382f5d8093cc871d156cf7294fee1f80b81c87c9ebbb1dff49490b230 Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.429488 4936 generic.go:334] "Generic (PLEG): container finished" podID="df4dcd90-70c5-4d83-b8a9-db6d914a8a2d" containerID="0b6794161057d16f86c7ab3b7cfa51d5923f92378407959f1d03bfd9c2857f77" exitCode=0 Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.429853 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7437-account-create-update-qn4hp" event={"ID":"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d","Type":"ContainerDied","Data":"0b6794161057d16f86c7ab3b7cfa51d5923f92378407959f1d03bfd9c2857f77"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.429883 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7437-account-create-update-qn4hp" event={"ID":"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d","Type":"ContainerStarted","Data":"bfad5c0970e7d2bfd871da507e2cfa365c8e2f7537f9562c6b74be94c4f3e2be"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.434416 4936 generic.go:334] "Generic (PLEG): container finished" podID="da80eafb-7c87-4b87-89ae-beee42d9b256" containerID="89fdffa55aba7f23c2c414ea31524a469dd7b82ebff0488389910f625da859c4" exitCode=0 Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.434490 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m5pnp" event={"ID":"da80eafb-7c87-4b87-89ae-beee42d9b256","Type":"ContainerDied","Data":"89fdffa55aba7f23c2c414ea31524a469dd7b82ebff0488389910f625da859c4"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.434522 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m5pnp" event={"ID":"da80eafb-7c87-4b87-89ae-beee42d9b256","Type":"ContainerStarted","Data":"2223c364c394b75a6d929bae217686701d40a0a9fb81e9ba0b81195e7f673a34"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.435961 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wfxxc" event={"ID":"e13cc3e4-9442-40ec-b43f-cb968e26a0e1","Type":"ContainerStarted","Data":"c97a6d75893ef854e9ad2203c5b5890b792b4145066823f5f2c6e9e8493b595e"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.438091 4936 generic.go:334] "Generic (PLEG): container finished" podID="df8f8d6e-b8f0-4148-940e-712399f97ab1" containerID="29a8dda96b57bda5eebb47e6d6bc02216ba11a1b882a776df41be2834c8291e0" exitCode=0 Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.438136 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-63a4-account-create-update-wng8g" event={"ID":"df8f8d6e-b8f0-4148-940e-712399f97ab1","Type":"ContainerDied","Data":"29a8dda96b57bda5eebb47e6d6bc02216ba11a1b882a776df41be2834c8291e0"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.438151 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-63a4-account-create-update-wng8g" event={"ID":"df8f8d6e-b8f0-4148-940e-712399f97ab1","Type":"ContainerStarted","Data":"bbfb0b56d496820847db1d7c275239ba920374acfa48227082d2cbb674536a24"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.439894 4936 generic.go:334] "Generic (PLEG): container finished" podID="b802c0e2-15be-4604-b64c-3b7441a1096d" containerID="54a706b4981c4361f3f965a7f06842c315d8cf05db79fa68b6a6fb3dc232c85a" exitCode=0 Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.439959 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ab53-account-create-update-rrk7z" event={"ID":"b802c0e2-15be-4604-b64c-3b7441a1096d","Type":"ContainerDied","Data":"54a706b4981c4361f3f965a7f06842c315d8cf05db79fa68b6a6fb3dc232c85a"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.439976 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ab53-account-create-update-rrk7z" event={"ID":"b802c0e2-15be-4604-b64c-3b7441a1096d","Type":"ContainerStarted","Data":"fff3d24a22d6f4da7829c84893c68df9e7c77e30383c0ef0d1589849166bb8c8"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.441672 4936 generic.go:334] "Generic (PLEG): container finished" podID="0288b360-6ff9-4568-bc31-1351a3e55ad2" containerID="3aa99a195b752614ecace8abbec25d0c2e4d058279b0d984fecc8d8198321f9c" exitCode=0 Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.441904 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s46fs" event={"ID":"0288b360-6ff9-4568-bc31-1351a3e55ad2","Type":"ContainerDied","Data":"3aa99a195b752614ecace8abbec25d0c2e4d058279b0d984fecc8d8198321f9c"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.441928 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s46fs" event={"ID":"0288b360-6ff9-4568-bc31-1351a3e55ad2","Type":"ContainerStarted","Data":"bc1638d382f5d8093cc871d156cf7294fee1f80b81c87c9ebbb1dff49490b230"} Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.854302 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-p249h" Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.987976 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2726b8ff-699c-419e-ae52-0868c355ea23-operator-scripts\") pod \"2726b8ff-699c-419e-ae52-0868c355ea23\" (UID: \"2726b8ff-699c-419e-ae52-0868c355ea23\") " Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.988063 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4dh8\" (UniqueName: \"kubernetes.io/projected/2726b8ff-699c-419e-ae52-0868c355ea23-kube-api-access-l4dh8\") pod \"2726b8ff-699c-419e-ae52-0868c355ea23\" (UID: \"2726b8ff-699c-419e-ae52-0868c355ea23\") " Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.989669 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2726b8ff-699c-419e-ae52-0868c355ea23-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2726b8ff-699c-419e-ae52-0868c355ea23" (UID: "2726b8ff-699c-419e-ae52-0868c355ea23"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:50 crc kubenswrapper[4936]: I0320 16:18:50.995395 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2726b8ff-699c-419e-ae52-0868c355ea23-kube-api-access-l4dh8" (OuterVolumeSpecName: "kube-api-access-l4dh8") pod "2726b8ff-699c-419e-ae52-0868c355ea23" (UID: "2726b8ff-699c-419e-ae52-0868c355ea23"). InnerVolumeSpecName "kube-api-access-l4dh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:51 crc kubenswrapper[4936]: I0320 16:18:51.089969 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2726b8ff-699c-419e-ae52-0868c355ea23-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:51 crc kubenswrapper[4936]: I0320 16:18:51.090012 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4dh8\" (UniqueName: \"kubernetes.io/projected/2726b8ff-699c-419e-ae52-0868c355ea23-kube-api-access-l4dh8\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:51 crc kubenswrapper[4936]: I0320 16:18:51.453306 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-p249h" event={"ID":"2726b8ff-699c-419e-ae52-0868c355ea23","Type":"ContainerDied","Data":"c5035aec792808e37a45d18ecf0672ad339bafa9b403c12682a47dd0d3179748"} Mar 20 16:18:51 crc kubenswrapper[4936]: I0320 16:18:51.453574 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5035aec792808e37a45d18ecf0672ad339bafa9b403c12682a47dd0d3179748" Mar 20 16:18:51 crc kubenswrapper[4936]: I0320 16:18:51.453441 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-p249h" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.408871 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.417448 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.452867 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.455316 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwcwq\" (UniqueName: \"kubernetes.io/projected/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-kube-api-access-nwcwq\") pod \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\" (UID: \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.455373 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-operator-scripts\") pod \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\" (UID: \"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.456191 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "df4dcd90-70c5-4d83-b8a9-db6d914a8a2d" (UID: "df4dcd90-70c5-4d83-b8a9-db6d914a8a2d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.456254 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.463533 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-kube-api-access-nwcwq" (OuterVolumeSpecName: "kube-api-access-nwcwq") pod "df4dcd90-70c5-4d83-b8a9-db6d914a8a2d" (UID: "df4dcd90-70c5-4d83-b8a9-db6d914a8a2d"). InnerVolumeSpecName "kube-api-access-nwcwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.475953 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.490926 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ab53-account-create-update-rrk7z" event={"ID":"b802c0e2-15be-4604-b64c-3b7441a1096d","Type":"ContainerDied","Data":"fff3d24a22d6f4da7829c84893c68df9e7c77e30383c0ef0d1589849166bb8c8"} Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.490971 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fff3d24a22d6f4da7829c84893c68df9e7c77e30383c0ef0d1589849166bb8c8" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.490935 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab53-account-create-update-rrk7z" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.492643 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s46fs" event={"ID":"0288b360-6ff9-4568-bc31-1351a3e55ad2","Type":"ContainerDied","Data":"bc1638d382f5d8093cc871d156cf7294fee1f80b81c87c9ebbb1dff49490b230"} Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.492669 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc1638d382f5d8093cc871d156cf7294fee1f80b81c87c9ebbb1dff49490b230" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.492724 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s46fs" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.494476 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7437-account-create-update-qn4hp" event={"ID":"df4dcd90-70c5-4d83-b8a9-db6d914a8a2d","Type":"ContainerDied","Data":"bfad5c0970e7d2bfd871da507e2cfa365c8e2f7537f9562c6b74be94c4f3e2be"} Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.494500 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfad5c0970e7d2bfd871da507e2cfa365c8e2f7537f9562c6b74be94c4f3e2be" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.494533 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7437-account-create-update-qn4hp" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.508207 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m5pnp" event={"ID":"da80eafb-7c87-4b87-89ae-beee42d9b256","Type":"ContainerDied","Data":"2223c364c394b75a6d929bae217686701d40a0a9fb81e9ba0b81195e7f673a34"} Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.508248 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2223c364c394b75a6d929bae217686701d40a0a9fb81e9ba0b81195e7f673a34" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.508225 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m5pnp" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.510940 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-63a4-account-create-update-wng8g" event={"ID":"df8f8d6e-b8f0-4148-940e-712399f97ab1","Type":"ContainerDied","Data":"bbfb0b56d496820847db1d7c275239ba920374acfa48227082d2cbb674536a24"} Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.510969 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-63a4-account-create-update-wng8g" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.510983 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbfb0b56d496820847db1d7c275239ba920374acfa48227082d2cbb674536a24" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.557051 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0288b360-6ff9-4568-bc31-1351a3e55ad2-operator-scripts\") pod \"0288b360-6ff9-4568-bc31-1351a3e55ad2\" (UID: \"0288b360-6ff9-4568-bc31-1351a3e55ad2\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.557506 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0288b360-6ff9-4568-bc31-1351a3e55ad2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0288b360-6ff9-4568-bc31-1351a3e55ad2" (UID: "0288b360-6ff9-4568-bc31-1351a3e55ad2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.557667 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df8f8d6e-b8f0-4148-940e-712399f97ab1-operator-scripts\") pod \"df8f8d6e-b8f0-4148-940e-712399f97ab1\" (UID: \"df8f8d6e-b8f0-4148-940e-712399f97ab1\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.558115 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df8f8d6e-b8f0-4148-940e-712399f97ab1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "df8f8d6e-b8f0-4148-940e-712399f97ab1" (UID: "df8f8d6e-b8f0-4148-940e-712399f97ab1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.558198 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s85s\" (UniqueName: \"kubernetes.io/projected/df8f8d6e-b8f0-4148-940e-712399f97ab1-kube-api-access-2s85s\") pod \"df8f8d6e-b8f0-4148-940e-712399f97ab1\" (UID: \"df8f8d6e-b8f0-4148-940e-712399f97ab1\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.558620 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbp4c\" (UniqueName: \"kubernetes.io/projected/b802c0e2-15be-4604-b64c-3b7441a1096d-kube-api-access-jbp4c\") pod \"b802c0e2-15be-4604-b64c-3b7441a1096d\" (UID: \"b802c0e2-15be-4604-b64c-3b7441a1096d\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.558648 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b802c0e2-15be-4604-b64c-3b7441a1096d-operator-scripts\") pod \"b802c0e2-15be-4604-b64c-3b7441a1096d\" (UID: \"b802c0e2-15be-4604-b64c-3b7441a1096d\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.558696 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gtnx\" (UniqueName: \"kubernetes.io/projected/da80eafb-7c87-4b87-89ae-beee42d9b256-kube-api-access-8gtnx\") pod \"da80eafb-7c87-4b87-89ae-beee42d9b256\" (UID: \"da80eafb-7c87-4b87-89ae-beee42d9b256\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.558724 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kx6m\" (UniqueName: \"kubernetes.io/projected/0288b360-6ff9-4568-bc31-1351a3e55ad2-kube-api-access-8kx6m\") pod \"0288b360-6ff9-4568-bc31-1351a3e55ad2\" (UID: \"0288b360-6ff9-4568-bc31-1351a3e55ad2\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.558772 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da80eafb-7c87-4b87-89ae-beee42d9b256-operator-scripts\") pod \"da80eafb-7c87-4b87-89ae-beee42d9b256\" (UID: \"da80eafb-7c87-4b87-89ae-beee42d9b256\") " Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.559189 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b802c0e2-15be-4604-b64c-3b7441a1096d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b802c0e2-15be-4604-b64c-3b7441a1096d" (UID: "b802c0e2-15be-4604-b64c-3b7441a1096d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.559347 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da80eafb-7c87-4b87-89ae-beee42d9b256-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "da80eafb-7c87-4b87-89ae-beee42d9b256" (UID: "da80eafb-7c87-4b87-89ae-beee42d9b256"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.559482 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df8f8d6e-b8f0-4148-940e-712399f97ab1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.559495 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b802c0e2-15be-4604-b64c-3b7441a1096d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.559504 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwcwq\" (UniqueName: \"kubernetes.io/projected/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-kube-api-access-nwcwq\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.559514 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.559523 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0288b360-6ff9-4568-bc31-1351a3e55ad2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.561887 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df8f8d6e-b8f0-4148-940e-712399f97ab1-kube-api-access-2s85s" (OuterVolumeSpecName: "kube-api-access-2s85s") pod "df8f8d6e-b8f0-4148-940e-712399f97ab1" (UID: "df8f8d6e-b8f0-4148-940e-712399f97ab1"). InnerVolumeSpecName "kube-api-access-2s85s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.563240 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0288b360-6ff9-4568-bc31-1351a3e55ad2-kube-api-access-8kx6m" (OuterVolumeSpecName: "kube-api-access-8kx6m") pod "0288b360-6ff9-4568-bc31-1351a3e55ad2" (UID: "0288b360-6ff9-4568-bc31-1351a3e55ad2"). InnerVolumeSpecName "kube-api-access-8kx6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.563331 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da80eafb-7c87-4b87-89ae-beee42d9b256-kube-api-access-8gtnx" (OuterVolumeSpecName: "kube-api-access-8gtnx") pod "da80eafb-7c87-4b87-89ae-beee42d9b256" (UID: "da80eafb-7c87-4b87-89ae-beee42d9b256"). InnerVolumeSpecName "kube-api-access-8gtnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.563377 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b802c0e2-15be-4604-b64c-3b7441a1096d-kube-api-access-jbp4c" (OuterVolumeSpecName: "kube-api-access-jbp4c") pod "b802c0e2-15be-4604-b64c-3b7441a1096d" (UID: "b802c0e2-15be-4604-b64c-3b7441a1096d"). InnerVolumeSpecName "kube-api-access-jbp4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.661158 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s85s\" (UniqueName: \"kubernetes.io/projected/df8f8d6e-b8f0-4148-940e-712399f97ab1-kube-api-access-2s85s\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.661190 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbp4c\" (UniqueName: \"kubernetes.io/projected/b802c0e2-15be-4604-b64c-3b7441a1096d-kube-api-access-jbp4c\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.661199 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gtnx\" (UniqueName: \"kubernetes.io/projected/da80eafb-7c87-4b87-89ae-beee42d9b256-kube-api-access-8gtnx\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.661209 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kx6m\" (UniqueName: \"kubernetes.io/projected/0288b360-6ff9-4568-bc31-1351a3e55ad2-kube-api-access-8kx6m\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:54 crc kubenswrapper[4936]: I0320 16:18:54.661219 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da80eafb-7c87-4b87-89ae-beee42d9b256-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:55 crc kubenswrapper[4936]: I0320 16:18:55.536708 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wfxxc" event={"ID":"e13cc3e4-9442-40ec-b43f-cb968e26a0e1","Type":"ContainerStarted","Data":"e2964864b07f763a8cdc75e13ac493e4e214e805a94f28ce5767295b5474e76b"} Mar 20 16:18:55 crc kubenswrapper[4936]: I0320 16:18:55.565943 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wfxxc" podStartSLOduration=2.933522729 podStartE2EDuration="7.565924857s" podCreationTimestamp="2026-03-20 16:18:48 +0000 UTC" firstStartedPulling="2026-03-20 16:18:49.628418667 +0000 UTC m=+1080.574786482" lastFinishedPulling="2026-03-20 16:18:54.260820795 +0000 UTC m=+1085.207188610" observedRunningTime="2026-03-20 16:18:55.557944055 +0000 UTC m=+1086.504311870" watchObservedRunningTime="2026-03-20 16:18:55.565924857 +0000 UTC m=+1086.512292672" Mar 20 16:18:57 crc kubenswrapper[4936]: I0320 16:18:57.555848 4936 generic.go:334] "Generic (PLEG): container finished" podID="e13cc3e4-9442-40ec-b43f-cb968e26a0e1" containerID="e2964864b07f763a8cdc75e13ac493e4e214e805a94f28ce5767295b5474e76b" exitCode=0 Mar 20 16:18:57 crc kubenswrapper[4936]: I0320 16:18:57.555941 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wfxxc" event={"ID":"e13cc3e4-9442-40ec-b43f-cb968e26a0e1","Type":"ContainerDied","Data":"e2964864b07f763a8cdc75e13ac493e4e214e805a94f28ce5767295b5474e76b"} Mar 20 16:18:58 crc kubenswrapper[4936]: I0320 16:18:58.854793 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:18:58 crc kubenswrapper[4936]: I0320 16:18:58.855061 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:18:58 crc kubenswrapper[4936]: I0320 16:18:58.940386 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.033488 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-combined-ca-bundle\") pod \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.033892 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb7qk\" (UniqueName: \"kubernetes.io/projected/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-kube-api-access-cb7qk\") pod \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.034009 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-config-data\") pod \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\" (UID: \"e13cc3e4-9442-40ec-b43f-cb968e26a0e1\") " Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.045887 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-kube-api-access-cb7qk" (OuterVolumeSpecName: "kube-api-access-cb7qk") pod "e13cc3e4-9442-40ec-b43f-cb968e26a0e1" (UID: "e13cc3e4-9442-40ec-b43f-cb968e26a0e1"). InnerVolumeSpecName "kube-api-access-cb7qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.061034 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e13cc3e4-9442-40ec-b43f-cb968e26a0e1" (UID: "e13cc3e4-9442-40ec-b43f-cb968e26a0e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.090780 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-config-data" (OuterVolumeSpecName: "config-data") pod "e13cc3e4-9442-40ec-b43f-cb968e26a0e1" (UID: "e13cc3e4-9442-40ec-b43f-cb968e26a0e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.135644 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb7qk\" (UniqueName: \"kubernetes.io/projected/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-kube-api-access-cb7qk\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.135681 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.135692 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13cc3e4-9442-40ec-b43f-cb968e26a0e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.576477 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wfxxc" event={"ID":"e13cc3e4-9442-40ec-b43f-cb968e26a0e1","Type":"ContainerDied","Data":"c97a6d75893ef854e9ad2203c5b5890b792b4145066823f5f2c6e9e8493b595e"} Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.576778 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c97a6d75893ef854e9ad2203c5b5890b792b4145066823f5f2c6e9e8493b595e" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.576729 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wfxxc" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880229 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-2bz9f"] Mar 20 16:18:59 crc kubenswrapper[4936]: E0320 16:18:59.880632 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0288b360-6ff9-4568-bc31-1351a3e55ad2" containerName="mariadb-database-create" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880651 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="0288b360-6ff9-4568-bc31-1351a3e55ad2" containerName="mariadb-database-create" Mar 20 16:18:59 crc kubenswrapper[4936]: E0320 16:18:59.880665 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4dcd90-70c5-4d83-b8a9-db6d914a8a2d" containerName="mariadb-account-create-update" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880673 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4dcd90-70c5-4d83-b8a9-db6d914a8a2d" containerName="mariadb-account-create-update" Mar 20 16:18:59 crc kubenswrapper[4936]: E0320 16:18:59.880682 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da80eafb-7c87-4b87-89ae-beee42d9b256" containerName="mariadb-database-create" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880689 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="da80eafb-7c87-4b87-89ae-beee42d9b256" containerName="mariadb-database-create" Mar 20 16:18:59 crc kubenswrapper[4936]: E0320 16:18:59.880702 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b802c0e2-15be-4604-b64c-3b7441a1096d" containerName="mariadb-account-create-update" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880708 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b802c0e2-15be-4604-b64c-3b7441a1096d" containerName="mariadb-account-create-update" Mar 20 16:18:59 crc kubenswrapper[4936]: E0320 16:18:59.880728 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df8f8d6e-b8f0-4148-940e-712399f97ab1" containerName="mariadb-account-create-update" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880735 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="df8f8d6e-b8f0-4148-940e-712399f97ab1" containerName="mariadb-account-create-update" Mar 20 16:18:59 crc kubenswrapper[4936]: E0320 16:18:59.880747 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2726b8ff-699c-419e-ae52-0868c355ea23" containerName="mariadb-database-create" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880755 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="2726b8ff-699c-419e-ae52-0868c355ea23" containerName="mariadb-database-create" Mar 20 16:18:59 crc kubenswrapper[4936]: E0320 16:18:59.880772 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13cc3e4-9442-40ec-b43f-cb968e26a0e1" containerName="keystone-db-sync" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880780 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13cc3e4-9442-40ec-b43f-cb968e26a0e1" containerName="keystone-db-sync" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880965 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="0288b360-6ff9-4568-bc31-1351a3e55ad2" containerName="mariadb-database-create" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880984 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="df4dcd90-70c5-4d83-b8a9-db6d914a8a2d" containerName="mariadb-account-create-update" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.880993 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13cc3e4-9442-40ec-b43f-cb968e26a0e1" containerName="keystone-db-sync" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.881010 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="b802c0e2-15be-4604-b64c-3b7441a1096d" containerName="mariadb-account-create-update" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.881020 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="2726b8ff-699c-419e-ae52-0868c355ea23" containerName="mariadb-database-create" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.881032 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="df8f8d6e-b8f0-4148-940e-712399f97ab1" containerName="mariadb-account-create-update" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.881049 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="da80eafb-7c87-4b87-89ae-beee42d9b256" containerName="mariadb-database-create" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.882668 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.896413 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-2bz9f"] Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.947309 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-qmlw5"] Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.950135 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.953930 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.954258 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.954821 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gbshp" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.954901 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.955121 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 20 16:18:59 crc kubenswrapper[4936]: I0320 16:18:59.955350 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qmlw5"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.048680 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.048846 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsqfz\" (UniqueName: \"kubernetes.io/projected/5833ca82-5b16-49a3-be48-1a0c0ab7274f-kube-api-access-hsqfz\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.048980 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.049262 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.049302 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-svc\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.049344 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-config\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.150947 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-svc\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.150998 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-combined-ca-bundle\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151023 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-config\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151057 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-credential-keys\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151095 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151128 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsqfz\" (UniqueName: \"kubernetes.io/projected/5833ca82-5b16-49a3-be48-1a0c0ab7274f-kube-api-access-hsqfz\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151186 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151232 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-fernet-keys\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151265 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pxf4\" (UniqueName: \"kubernetes.io/projected/62abfd11-7276-46a9-b53a-9e90fc8ebbde-kube-api-access-6pxf4\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151296 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-scripts\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151317 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-config-data\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.151340 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.152087 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.152127 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.152603 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-svc\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.153095 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.153103 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-config\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.175585 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-n7kq7"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.185741 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.192790 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.193123 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.193288 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7kz5m" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.194406 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsqfz\" (UniqueName: \"kubernetes.io/projected/5833ca82-5b16-49a3-be48-1a0c0ab7274f-kube-api-access-hsqfz\") pod \"dnsmasq-dns-847c4cc679-2bz9f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.201019 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.204706 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74b7b699cf-b6cr9"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.205973 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.207496 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.208465 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.216844 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.216844 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-7ptxp" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.222812 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-n7kq7"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.262609 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pxf4\" (UniqueName: \"kubernetes.io/projected/62abfd11-7276-46a9-b53a-9e90fc8ebbde-kube-api-access-6pxf4\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.262967 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-scripts\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.262996 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-config-data\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.263033 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-combined-ca-bundle\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.263078 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-credential-keys\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.263177 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-fernet-keys\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.265944 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74b7b699cf-b6cr9"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.273967 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-config-data\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.293665 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-scripts\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.296334 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-fernet-keys\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.297533 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-credential-keys\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.313186 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-combined-ca-bundle\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.329229 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pxf4\" (UniqueName: \"kubernetes.io/projected/62abfd11-7276-46a9-b53a-9e90fc8ebbde-kube-api-access-6pxf4\") pod \"keystone-bootstrap-qmlw5\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.346440 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4t65w"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.347923 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.354063 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-txrnm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.354332 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.354453 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.364240 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-2bz9f"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365154 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/047fb63e-64a8-4280-9846-659575038df3-etc-machine-id\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365194 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-config-data\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365216 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7f8b999-b632-4354-bc26-925b418b379e-logs\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365264 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-combined-ca-bundle\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365282 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-scripts\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365307 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-scripts\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365335 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd9wf\" (UniqueName: \"kubernetes.io/projected/b7f8b999-b632-4354-bc26-925b418b379e-kube-api-access-jd9wf\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365358 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b7f8b999-b632-4354-bc26-925b418b379e-horizon-secret-key\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365383 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j49gd\" (UniqueName: \"kubernetes.io/projected/047fb63e-64a8-4280-9846-659575038df3-kube-api-access-j49gd\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365407 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-db-sync-config-data\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.365427 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-config-data\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.378613 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9w2gg"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.379850 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.388485 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.388677 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rpkn6" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.399389 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4t65w"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.425712 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-j4d5s"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.427076 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.437809 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.437929 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qmxxr" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.437998 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.442489 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-j4d5s"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.487793 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-scripts\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.488079 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-scripts\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.489676 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-config-data\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.489803 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd9wf\" (UniqueName: \"kubernetes.io/projected/b7f8b999-b632-4354-bc26-925b418b379e-kube-api-access-jd9wf\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.489930 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b7f8b999-b632-4354-bc26-925b418b379e-horizon-secret-key\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.490018 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b87c4418-b13e-41e2-bbd8-d7dfb057a594-logs\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.490169 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-combined-ca-bundle\") pod \"barbican-db-sync-9w2gg\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.490254 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g497k\" (UniqueName: \"kubernetes.io/projected/b87c4418-b13e-41e2-bbd8-d7dfb057a594-kube-api-access-g497k\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.491327 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j49gd\" (UniqueName: \"kubernetes.io/projected/047fb63e-64a8-4280-9846-659575038df3-kube-api-access-j49gd\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.491463 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-db-sync-config-data\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.491584 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85cwt\" (UniqueName: \"kubernetes.io/projected/61aff63e-dc1b-4854-af4a-188837ad4cf9-kube-api-access-85cwt\") pod \"barbican-db-sync-9w2gg\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.491713 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-config-data\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.492733 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-db-sync-config-data\") pod \"barbican-db-sync-9w2gg\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.492876 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/047fb63e-64a8-4280-9846-659575038df3-etc-machine-id\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.493005 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-combined-ca-bundle\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.493101 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-config-data\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.493186 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7f8b999-b632-4354-bc26-925b418b379e-logs\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.493406 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-config\") pod \"neutron-db-sync-4t65w\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.493582 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw2x5\" (UniqueName: \"kubernetes.io/projected/5421152c-d474-428f-9554-96e036719655-kube-api-access-rw2x5\") pod \"neutron-db-sync-4t65w\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.493718 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-combined-ca-bundle\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.493817 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-scripts\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.493955 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-combined-ca-bundle\") pod \"neutron-db-sync-4t65w\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.504455 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-config-data\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.507195 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-db-sync-config-data\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.509208 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7f8b999-b632-4354-bc26-925b418b379e-logs\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.511699 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/047fb63e-64a8-4280-9846-659575038df3-etc-machine-id\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.512409 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-scripts\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.512928 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-config-data\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.528162 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b7f8b999-b632-4354-bc26-925b418b379e-horizon-secret-key\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.532180 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-combined-ca-bundle\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.545576 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd9wf\" (UniqueName: \"kubernetes.io/projected/b7f8b999-b632-4354-bc26-925b418b379e-kube-api-access-jd9wf\") pod \"horizon-74b7b699cf-b6cr9\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.554495 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9w2gg"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.574644 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-bnzqf"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.581298 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j49gd\" (UniqueName: \"kubernetes.io/projected/047fb63e-64a8-4280-9846-659575038df3-kube-api-access-j49gd\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.581624 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.591887 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600062 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-scripts\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600105 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-config-data\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600139 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b87c4418-b13e-41e2-bbd8-d7dfb057a594-logs\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600167 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-combined-ca-bundle\") pod \"barbican-db-sync-9w2gg\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600193 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g497k\" (UniqueName: \"kubernetes.io/projected/b87c4418-b13e-41e2-bbd8-d7dfb057a594-kube-api-access-g497k\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600232 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85cwt\" (UniqueName: \"kubernetes.io/projected/61aff63e-dc1b-4854-af4a-188837ad4cf9-kube-api-access-85cwt\") pod \"barbican-db-sync-9w2gg\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600264 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-db-sync-config-data\") pod \"barbican-db-sync-9w2gg\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600298 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-combined-ca-bundle\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600350 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-config\") pod \"neutron-db-sync-4t65w\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600378 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw2x5\" (UniqueName: \"kubernetes.io/projected/5421152c-d474-428f-9554-96e036719655-kube-api-access-rw2x5\") pod \"neutron-db-sync-4t65w\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.600404 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-combined-ca-bundle\") pod \"neutron-db-sync-4t65w\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.606943 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b87c4418-b13e-41e2-bbd8-d7dfb057a594-logs\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.612144 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-scripts\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.636959 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-db-sync-config-data\") pod \"barbican-db-sync-9w2gg\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.647841 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-config-data\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.652446 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-scripts\") pod \"cinder-db-sync-n7kq7\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.653689 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-config\") pod \"neutron-db-sync-4t65w\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.657226 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-combined-ca-bundle\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.658312 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-678fc86c4f-mxcbm"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.660189 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85cwt\" (UniqueName: \"kubernetes.io/projected/61aff63e-dc1b-4854-af4a-188837ad4cf9-kube-api-access-85cwt\") pod \"barbican-db-sync-9w2gg\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.660803 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw2x5\" (UniqueName: \"kubernetes.io/projected/5421152c-d474-428f-9554-96e036719655-kube-api-access-rw2x5\") pod \"neutron-db-sync-4t65w\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.665791 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.667013 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g497k\" (UniqueName: \"kubernetes.io/projected/b87c4418-b13e-41e2-bbd8-d7dfb057a594-kube-api-access-g497k\") pod \"placement-db-sync-j4d5s\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.669991 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-combined-ca-bundle\") pod \"neutron-db-sync-4t65w\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.690030 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-combined-ca-bundle\") pod \"barbican-db-sync-9w2gg\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.706736 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.706814 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.706883 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-config\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.706925 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.708699 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.708803 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.709280 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhgx4\" (UniqueName: \"kubernetes.io/projected/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-kube-api-access-bhgx4\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.719198 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-678fc86c4f-mxcbm"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.720684 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.739169 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-bnzqf"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.744084 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.773218 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.787278 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.789347 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.792798 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.799618 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.810508 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4c49318-0b4d-48d2-aeb1-49e74d760e82-horizon-secret-key\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.810700 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhgx4\" (UniqueName: \"kubernetes.io/projected/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-kube-api-access-bhgx4\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.810775 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4c49318-0b4d-48d2-aeb1-49e74d760e82-logs\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.810803 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.810842 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.810891 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-config\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.810925 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.810957 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kht7\" (UniqueName: \"kubernetes.io/projected/f4c49318-0b4d-48d2-aeb1-49e74d760e82-kube-api-access-2kht7\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.810983 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-scripts\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.811021 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.811053 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-config-data\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.812372 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.813880 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.814891 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-config\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.815401 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.815486 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.827639 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.834288 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhgx4\" (UniqueName: \"kubernetes.io/projected/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-kube-api-access-bhgx4\") pod \"dnsmasq-dns-785d8bcb8c-bnzqf\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.844878 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.850066 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.853209 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.853491 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.853670 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.853902 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-l2w6m" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.855612 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j4d5s" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.861613 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.872661 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.878249 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.882356 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.882615 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.912228 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-scripts\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.912497 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.912650 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-config-data\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.912765 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4c49318-0b4d-48d2-aeb1-49e74d760e82-horizon-secret-key\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.912963 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4c49318-0b4d-48d2-aeb1-49e74d760e82-logs\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.913062 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.913165 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-scripts\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.913304 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-log-httpd\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.913379 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-scripts\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.913522 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrdhw\" (UniqueName: \"kubernetes.io/projected/052101db-ddd5-419e-8131-fccbe44241b4-kube-api-access-wrdhw\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.913758 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kht7\" (UniqueName: \"kubernetes.io/projected/f4c49318-0b4d-48d2-aeb1-49e74d760e82-kube-api-access-2kht7\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.914702 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-config-data\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.914831 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-run-httpd\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.914583 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4c49318-0b4d-48d2-aeb1-49e74d760e82-logs\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.915156 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.918277 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-config-data\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.936064 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4c49318-0b4d-48d2-aeb1-49e74d760e82-horizon-secret-key\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:00 crc kubenswrapper[4936]: I0320 16:19:00.938273 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kht7\" (UniqueName: \"kubernetes.io/projected/f4c49318-0b4d-48d2-aeb1-49e74d760e82-kube-api-access-2kht7\") pod \"horizon-678fc86c4f-mxcbm\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.005428 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.021944 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.021997 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkspl\" (UniqueName: \"kubernetes.io/projected/468d0fa0-8842-4ff6-a36e-d62ed287850a-kube-api-access-xkspl\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022031 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-log-httpd\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022105 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022130 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrdhw\" (UniqueName: \"kubernetes.io/projected/052101db-ddd5-419e-8131-fccbe44241b4-kube-api-access-wrdhw\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022162 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-config-data\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022184 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-config-data\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022207 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022227 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-run-httpd\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022250 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022281 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022312 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022383 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxxzc\" (UniqueName: \"kubernetes.io/projected/5233e4bc-903d-4c6a-8319-085a365268d0-kube-api-access-pxxzc\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022415 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-logs\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022440 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022469 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-logs\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022515 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-scripts\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022583 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022627 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022648 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022674 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022705 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-scripts\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.022731 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.023286 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.024891 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-run-httpd\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.026015 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-log-httpd\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.034157 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.035666 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-scripts\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.036794 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-config-data\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.047260 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.047744 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrdhw\" (UniqueName: \"kubernetes.io/projected/052101db-ddd5-419e-8131-fccbe44241b4-kube-api-access-wrdhw\") pod \"ceilometer-0\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.098015 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-2bz9f"] Mar 20 16:19:01 crc kubenswrapper[4936]: W0320 16:19:01.104135 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5833ca82_5b16_49a3_be48_1a0c0ab7274f.slice/crio-42187260461bffdd2eb31aa75c07cfe14eeb6a43a73efa9f135bf3077cce6bbb WatchSource:0}: Error finding container 42187260461bffdd2eb31aa75c07cfe14eeb6a43a73efa9f135bf3077cce6bbb: Status 404 returned error can't find the container with id 42187260461bffdd2eb31aa75c07cfe14eeb6a43a73efa9f135bf3077cce6bbb Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.124597 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.124782 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-config-data\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.124880 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.124951 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.125045 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.125181 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxxzc\" (UniqueName: \"kubernetes.io/projected/5233e4bc-903d-4c6a-8319-085a365268d0-kube-api-access-pxxzc\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.125283 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-logs\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.125373 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.125461 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-logs\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.125576 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-scripts\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.125687 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.125811 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.125921 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.126043 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.126176 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.126262 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkspl\" (UniqueName: \"kubernetes.io/projected/468d0fa0-8842-4ff6-a36e-d62ed287850a-kube-api-access-xkspl\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.131901 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.131922 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-config-data\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.132233 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-scripts\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.132702 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.134959 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.137835 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.140034 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-logs\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.140104 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.140139 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.140322 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-logs\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.141970 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.142587 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.145801 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.147197 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.151813 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.155894 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkspl\" (UniqueName: \"kubernetes.io/projected/468d0fa0-8842-4ff6-a36e-d62ed287850a-kube-api-access-xkspl\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.169705 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxxzc\" (UniqueName: \"kubernetes.io/projected/5233e4bc-903d-4c6a-8319-085a365268d0-kube-api-access-pxxzc\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.248818 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.253365 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.324856 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qmlw5"] Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.478017 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.538329 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.680035 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" event={"ID":"5833ca82-5b16-49a3-be48-1a0c0ab7274f","Type":"ContainerStarted","Data":"42187260461bffdd2eb31aa75c07cfe14eeb6a43a73efa9f135bf3077cce6bbb"} Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.696895 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-n7kq7"] Mar 20 16:19:01 crc kubenswrapper[4936]: I0320 16:19:01.703759 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qmlw5" event={"ID":"62abfd11-7276-46a9-b53a-9e90fc8ebbde","Type":"ContainerStarted","Data":"9baa07eb7b9f338df72ae7582429e707f6f75e85ed07400bf5591c93e3cbbe06"} Mar 20 16:19:01 crc kubenswrapper[4936]: W0320 16:19:01.771055 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod047fb63e_64a8_4280_9846_659575038df3.slice/crio-e8d55bc7390adcfd9d73f556db8df5783b0e984ca6753ed85c82967c7bace81d WatchSource:0}: Error finding container e8d55bc7390adcfd9d73f556db8df5783b0e984ca6753ed85c82967c7bace81d: Status 404 returned error can't find the container with id e8d55bc7390adcfd9d73f556db8df5783b0e984ca6753ed85c82967c7bace81d Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.136952 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-678fc86c4f-mxcbm"] Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.151585 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74b7b699cf-b6cr9"] Mar 20 16:19:02 crc kubenswrapper[4936]: W0320 16:19:02.154890 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4c49318_0b4d_48d2_aeb1_49e74d760e82.slice/crio-692ca8f29e2d6c576b1814ba84ef663ffafb77475156dc143c93fd6e335067c9 WatchSource:0}: Error finding container 692ca8f29e2d6c576b1814ba84ef663ffafb77475156dc143c93fd6e335067c9: Status 404 returned error can't find the container with id 692ca8f29e2d6c576b1814ba84ef663ffafb77475156dc143c93fd6e335067c9 Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.175090 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9w2gg"] Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.198132 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4t65w"] Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.310379 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-j4d5s"] Mar 20 16:19:02 crc kubenswrapper[4936]: W0320 16:19:02.329935 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb87c4418_b13e_41e2_bbd8_d7dfb057a594.slice/crio-241b564a85bfa65256f50f3c4577bf576121d53be0d0ab404e2483ed2f9b616c WatchSource:0}: Error finding container 241b564a85bfa65256f50f3c4577bf576121d53be0d0ab404e2483ed2f9b616c: Status 404 returned error can't find the container with id 241b564a85bfa65256f50f3c4577bf576121d53be0d0ab404e2483ed2f9b616c Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.331073 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.347635 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-bnzqf"] Mar 20 16:19:02 crc kubenswrapper[4936]: W0320 16:19:02.383921 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea9e43bd_bedb_48dd_ac3b_bea79379dba0.slice/crio-3900f92575c8232b772ffd98b70ce2de539b2b95db78a10e89004e5b92b4ee3d WatchSource:0}: Error finding container 3900f92575c8232b772ffd98b70ce2de539b2b95db78a10e89004e5b92b4ee3d: Status 404 returned error can't find the container with id 3900f92575c8232b772ffd98b70ce2de539b2b95db78a10e89004e5b92b4ee3d Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.581596 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:19:02 crc kubenswrapper[4936]: W0320 16:19:02.622481 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5233e4bc_903d_4c6a_8319_085a365268d0.slice/crio-56922ff6855d30eef029b7c5b5d9d04f0a821d2e785094f2716d6c3a6c8ae8ff WatchSource:0}: Error finding container 56922ff6855d30eef029b7c5b5d9d04f0a821d2e785094f2716d6c3a6c8ae8ff: Status 404 returned error can't find the container with id 56922ff6855d30eef029b7c5b5d9d04f0a821d2e785094f2716d6c3a6c8ae8ff Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.727068 4936 generic.go:334] "Generic (PLEG): container finished" podID="5833ca82-5b16-49a3-be48-1a0c0ab7274f" containerID="71ab1756970ce5150a91c93333f07c410e551aa5d1b66f903c37f9e440d7b850" exitCode=0 Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.727286 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" event={"ID":"5833ca82-5b16-49a3-be48-1a0c0ab7274f","Type":"ContainerDied","Data":"71ab1756970ce5150a91c93333f07c410e551aa5d1b66f903c37f9e440d7b850"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.732957 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9w2gg" event={"ID":"61aff63e-dc1b-4854-af4a-188837ad4cf9","Type":"ContainerStarted","Data":"3b77256931847b16a6ddf789a7af732fcc54c4c9b2fff549bc32f63d006f2ba1"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.745449 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4t65w" event={"ID":"5421152c-d474-428f-9554-96e036719655","Type":"ContainerStarted","Data":"7c4fc58febfccc4dbd4458de7a1dee9f847f11da75accafe6e2b829b70cd92f0"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.745571 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4t65w" event={"ID":"5421152c-d474-428f-9554-96e036719655","Type":"ContainerStarted","Data":"f8f86ce6dceba12a1119e48f55635df8ffc6e779d69448863acb2ddfa767147d"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.747106 4936 generic.go:334] "Generic (PLEG): container finished" podID="ea9e43bd-bedb-48dd-ac3b-bea79379dba0" containerID="636c1037b4c4e3eae5558e4631e0729053e7fb4b012927f4e120352c40cba7bf" exitCode=0 Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.747177 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" event={"ID":"ea9e43bd-bedb-48dd-ac3b-bea79379dba0","Type":"ContainerDied","Data":"636c1037b4c4e3eae5558e4631e0729053e7fb4b012927f4e120352c40cba7bf"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.747208 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" event={"ID":"ea9e43bd-bedb-48dd-ac3b-bea79379dba0","Type":"ContainerStarted","Data":"3900f92575c8232b772ffd98b70ce2de539b2b95db78a10e89004e5b92b4ee3d"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.777163 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qmlw5" event={"ID":"62abfd11-7276-46a9-b53a-9e90fc8ebbde","Type":"ContainerStarted","Data":"500981b2dbc9c7c4e6e72392c7242fa937724f289025c56305f488d06c177e0c"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.781740 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b7b699cf-b6cr9" event={"ID":"b7f8b999-b632-4354-bc26-925b418b379e","Type":"ContainerStarted","Data":"deedeb6f023a16f466a478a4a1010b30345eb3d2376f90b037f0879ceb59d814"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.808198 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"052101db-ddd5-419e-8131-fccbe44241b4","Type":"ContainerStarted","Data":"8a346d219c8a9854c8e7822f48f552381141090bce1735a9d41237dffbb7d512"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.811891 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5233e4bc-903d-4c6a-8319-085a365268d0","Type":"ContainerStarted","Data":"56922ff6855d30eef029b7c5b5d9d04f0a821d2e785094f2716d6c3a6c8ae8ff"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.814536 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n7kq7" event={"ID":"047fb63e-64a8-4280-9846-659575038df3","Type":"ContainerStarted","Data":"e8d55bc7390adcfd9d73f556db8df5783b0e984ca6753ed85c82967c7bace81d"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.840035 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678fc86c4f-mxcbm" event={"ID":"f4c49318-0b4d-48d2-aeb1-49e74d760e82","Type":"ContainerStarted","Data":"692ca8f29e2d6c576b1814ba84ef663ffafb77475156dc143c93fd6e335067c9"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.841912 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j4d5s" event={"ID":"b87c4418-b13e-41e2-bbd8-d7dfb057a594","Type":"ContainerStarted","Data":"241b564a85bfa65256f50f3c4577bf576121d53be0d0ab404e2483ed2f9b616c"} Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.848978 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4t65w" podStartSLOduration=2.84895822 podStartE2EDuration="2.84895822s" podCreationTimestamp="2026-03-20 16:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:02.805968647 +0000 UTC m=+1093.752336462" watchObservedRunningTime="2026-03-20 16:19:02.84895822 +0000 UTC m=+1093.795326035" Mar 20 16:19:02 crc kubenswrapper[4936]: W0320 16:19:02.866883 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod468d0fa0_8842_4ff6_a36e_d62ed287850a.slice/crio-3b6887c4ee7343b4c920ba6ebeaf9b8454c0229caa0c512483792220007bcb0b WatchSource:0}: Error finding container 3b6887c4ee7343b4c920ba6ebeaf9b8454c0229caa0c512483792220007bcb0b: Status 404 returned error can't find the container with id 3b6887c4ee7343b4c920ba6ebeaf9b8454c0229caa0c512483792220007bcb0b Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.872765 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:19:02 crc kubenswrapper[4936]: I0320 16:19:02.881492 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-qmlw5" podStartSLOduration=3.881468907 podStartE2EDuration="3.881468907s" podCreationTimestamp="2026-03-20 16:18:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:02.830686337 +0000 UTC m=+1093.777054172" watchObservedRunningTime="2026-03-20 16:19:02.881468907 +0000 UTC m=+1093.827836722" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.282184 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.344654 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-678fc86c4f-mxcbm"] Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.352463 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.380087 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-754c5bd75c-9hhl6"] Mar 20 16:19:03 crc kubenswrapper[4936]: E0320 16:19:03.380662 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5833ca82-5b16-49a3-be48-1a0c0ab7274f" containerName="init" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.380681 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="5833ca82-5b16-49a3-be48-1a0c0ab7274f" containerName="init" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.380884 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="5833ca82-5b16-49a3-be48-1a0c0ab7274f" containerName="init" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.381925 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.390869 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-754c5bd75c-9hhl6"] Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.431939 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.473021 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.527358 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-sb\") pod \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.527476 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-swift-storage-0\") pod \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.530636 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-nb\") pod \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.531372 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsqfz\" (UniqueName: \"kubernetes.io/projected/5833ca82-5b16-49a3-be48-1a0c0ab7274f-kube-api-access-hsqfz\") pod \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.532181 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-config\") pod \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.533038 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-svc\") pod \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\" (UID: \"5833ca82-5b16-49a3-be48-1a0c0ab7274f\") " Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.555996 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-config-data\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.556078 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19df092c-5039-4069-9010-16248fb51d14-logs\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.556253 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/19df092c-5039-4069-9010-16248fb51d14-horizon-secret-key\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.556377 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2h52\" (UniqueName: \"kubernetes.io/projected/19df092c-5039-4069-9010-16248fb51d14-kube-api-access-d2h52\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.556417 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-scripts\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.558141 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5833ca82-5b16-49a3-be48-1a0c0ab7274f-kube-api-access-hsqfz" (OuterVolumeSpecName: "kube-api-access-hsqfz") pod "5833ca82-5b16-49a3-be48-1a0c0ab7274f" (UID: "5833ca82-5b16-49a3-be48-1a0c0ab7274f"). InnerVolumeSpecName "kube-api-access-hsqfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.570072 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5833ca82-5b16-49a3-be48-1a0c0ab7274f" (UID: "5833ca82-5b16-49a3-be48-1a0c0ab7274f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.579153 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5833ca82-5b16-49a3-be48-1a0c0ab7274f" (UID: "5833ca82-5b16-49a3-be48-1a0c0ab7274f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.594861 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5833ca82-5b16-49a3-be48-1a0c0ab7274f" (UID: "5833ca82-5b16-49a3-be48-1a0c0ab7274f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.613068 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5833ca82-5b16-49a3-be48-1a0c0ab7274f" (UID: "5833ca82-5b16-49a3-be48-1a0c0ab7274f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.631415 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-config" (OuterVolumeSpecName: "config") pod "5833ca82-5b16-49a3-be48-1a0c0ab7274f" (UID: "5833ca82-5b16-49a3-be48-1a0c0ab7274f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658514 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-config-data\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658582 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19df092c-5039-4069-9010-16248fb51d14-logs\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658649 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/19df092c-5039-4069-9010-16248fb51d14-horizon-secret-key\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658713 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2h52\" (UniqueName: \"kubernetes.io/projected/19df092c-5039-4069-9010-16248fb51d14-kube-api-access-d2h52\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658745 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-scripts\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658827 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658839 4936 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658850 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658860 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsqfz\" (UniqueName: \"kubernetes.io/projected/5833ca82-5b16-49a3-be48-1a0c0ab7274f-kube-api-access-hsqfz\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658872 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.658882 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5833ca82-5b16-49a3-be48-1a0c0ab7274f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.659641 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-scripts\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.659712 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-config-data\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.659905 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19df092c-5039-4069-9010-16248fb51d14-logs\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.665376 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/19df092c-5039-4069-9010-16248fb51d14-horizon-secret-key\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.678701 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2h52\" (UniqueName: \"kubernetes.io/projected/19df092c-5039-4069-9010-16248fb51d14-kube-api-access-d2h52\") pod \"horizon-754c5bd75c-9hhl6\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.713609 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.867760 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.879514 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-2bz9f" event={"ID":"5833ca82-5b16-49a3-be48-1a0c0ab7274f","Type":"ContainerDied","Data":"42187260461bffdd2eb31aa75c07cfe14eeb6a43a73efa9f135bf3077cce6bbb"} Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.879596 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.879612 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" event={"ID":"ea9e43bd-bedb-48dd-ac3b-bea79379dba0","Type":"ContainerStarted","Data":"efa25416797404cf3c4d6b6637105ca24d3dfe58e17dfd86e05e9354d2c995a5"} Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.879624 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"468d0fa0-8842-4ff6-a36e-d62ed287850a","Type":"ContainerStarted","Data":"3b6887c4ee7343b4c920ba6ebeaf9b8454c0229caa0c512483792220007bcb0b"} Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.879730 4936 scope.go:117] "RemoveContainer" containerID="71ab1756970ce5150a91c93333f07c410e551aa5d1b66f903c37f9e440d7b850" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.892365 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" podStartSLOduration=3.892345406 podStartE2EDuration="3.892345406s" podCreationTimestamp="2026-03-20 16:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:03.887950578 +0000 UTC m=+1094.834318393" watchObservedRunningTime="2026-03-20 16:19:03.892345406 +0000 UTC m=+1094.838713221" Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.970410 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-2bz9f"] Mar 20 16:19:03 crc kubenswrapper[4936]: I0320 16:19:03.985715 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-2bz9f"] Mar 20 16:19:04 crc kubenswrapper[4936]: I0320 16:19:04.362000 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-754c5bd75c-9hhl6"] Mar 20 16:19:04 crc kubenswrapper[4936]: W0320 16:19:04.442805 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19df092c_5039_4069_9010_16248fb51d14.slice/crio-6c88fa74ccbe8c6e16335e3c79b0c3e6ca9792b04e3ae2fb25fdb773eb4007bb WatchSource:0}: Error finding container 6c88fa74ccbe8c6e16335e3c79b0c3e6ca9792b04e3ae2fb25fdb773eb4007bb: Status 404 returned error can't find the container with id 6c88fa74ccbe8c6e16335e3c79b0c3e6ca9792b04e3ae2fb25fdb773eb4007bb Mar 20 16:19:04 crc kubenswrapper[4936]: I0320 16:19:04.889150 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754c5bd75c-9hhl6" event={"ID":"19df092c-5039-4069-9010-16248fb51d14","Type":"ContainerStarted","Data":"6c88fa74ccbe8c6e16335e3c79b0c3e6ca9792b04e3ae2fb25fdb773eb4007bb"} Mar 20 16:19:04 crc kubenswrapper[4936]: I0320 16:19:04.905870 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"468d0fa0-8842-4ff6-a36e-d62ed287850a","Type":"ContainerStarted","Data":"0f700379c382adf0d0212cfd0a1f3e8211bc897cbefd37061acee29757acc695"} Mar 20 16:19:04 crc kubenswrapper[4936]: I0320 16:19:04.920119 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5233e4bc-903d-4c6a-8319-085a365268d0","Type":"ContainerStarted","Data":"634b46ddadc9a4c5e4f2c417a943beef8f06198aea131fb1dbec29ce4b1ca4d9"} Mar 20 16:19:05 crc kubenswrapper[4936]: I0320 16:19:05.871826 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5833ca82-5b16-49a3-be48-1a0c0ab7274f" path="/var/lib/kubelet/pods/5833ca82-5b16-49a3-be48-1a0c0ab7274f/volumes" Mar 20 16:19:05 crc kubenswrapper[4936]: I0320 16:19:05.938170 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"468d0fa0-8842-4ff6-a36e-d62ed287850a","Type":"ContainerStarted","Data":"72abbdb093c7163b29219d6209d62fdc24d07574bb34a2715412536edf686b7a"} Mar 20 16:19:05 crc kubenswrapper[4936]: I0320 16:19:05.938353 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerName="glance-log" containerID="cri-o://0f700379c382adf0d0212cfd0a1f3e8211bc897cbefd37061acee29757acc695" gracePeriod=30 Mar 20 16:19:05 crc kubenswrapper[4936]: I0320 16:19:05.938419 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerName="glance-httpd" containerID="cri-o://72abbdb093c7163b29219d6209d62fdc24d07574bb34a2715412536edf686b7a" gracePeriod=30 Mar 20 16:19:05 crc kubenswrapper[4936]: I0320 16:19:05.945464 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5233e4bc-903d-4c6a-8319-085a365268d0","Type":"ContainerStarted","Data":"f155cd461d0ae85ad1062b9959124d1122212d885cb667e8395b461af6953cdd"} Mar 20 16:19:05 crc kubenswrapper[4936]: I0320 16:19:05.945642 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5233e4bc-903d-4c6a-8319-085a365268d0" containerName="glance-log" containerID="cri-o://634b46ddadc9a4c5e4f2c417a943beef8f06198aea131fb1dbec29ce4b1ca4d9" gracePeriod=30 Mar 20 16:19:05 crc kubenswrapper[4936]: I0320 16:19:05.945757 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5233e4bc-903d-4c6a-8319-085a365268d0" containerName="glance-httpd" containerID="cri-o://f155cd461d0ae85ad1062b9959124d1122212d885cb667e8395b461af6953cdd" gracePeriod=30 Mar 20 16:19:05 crc kubenswrapper[4936]: I0320 16:19:05.990319 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.990293474 podStartE2EDuration="5.990293474s" podCreationTimestamp="2026-03-20 16:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:05.96404646 +0000 UTC m=+1096.910414265" watchObservedRunningTime="2026-03-20 16:19:05.990293474 +0000 UTC m=+1096.936661289" Mar 20 16:19:06 crc kubenswrapper[4936]: I0320 16:19:06.011045 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.011015788 podStartE2EDuration="6.011015788s" podCreationTimestamp="2026-03-20 16:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:05.991812729 +0000 UTC m=+1096.938180554" watchObservedRunningTime="2026-03-20 16:19:06.011015788 +0000 UTC m=+1096.957383603" Mar 20 16:19:07 crc kubenswrapper[4936]: I0320 16:19:07.006464 4936 generic.go:334] "Generic (PLEG): container finished" podID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerID="72abbdb093c7163b29219d6209d62fdc24d07574bb34a2715412536edf686b7a" exitCode=0 Mar 20 16:19:07 crc kubenswrapper[4936]: I0320 16:19:07.006713 4936 generic.go:334] "Generic (PLEG): container finished" podID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerID="0f700379c382adf0d0212cfd0a1f3e8211bc897cbefd37061acee29757acc695" exitCode=143 Mar 20 16:19:07 crc kubenswrapper[4936]: I0320 16:19:07.006591 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"468d0fa0-8842-4ff6-a36e-d62ed287850a","Type":"ContainerDied","Data":"72abbdb093c7163b29219d6209d62fdc24d07574bb34a2715412536edf686b7a"} Mar 20 16:19:07 crc kubenswrapper[4936]: I0320 16:19:07.006772 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"468d0fa0-8842-4ff6-a36e-d62ed287850a","Type":"ContainerDied","Data":"0f700379c382adf0d0212cfd0a1f3e8211bc897cbefd37061acee29757acc695"} Mar 20 16:19:07 crc kubenswrapper[4936]: I0320 16:19:07.010825 4936 generic.go:334] "Generic (PLEG): container finished" podID="5233e4bc-903d-4c6a-8319-085a365268d0" containerID="f155cd461d0ae85ad1062b9959124d1122212d885cb667e8395b461af6953cdd" exitCode=0 Mar 20 16:19:07 crc kubenswrapper[4936]: I0320 16:19:07.010856 4936 generic.go:334] "Generic (PLEG): container finished" podID="5233e4bc-903d-4c6a-8319-085a365268d0" containerID="634b46ddadc9a4c5e4f2c417a943beef8f06198aea131fb1dbec29ce4b1ca4d9" exitCode=143 Mar 20 16:19:07 crc kubenswrapper[4936]: I0320 16:19:07.010876 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5233e4bc-903d-4c6a-8319-085a365268d0","Type":"ContainerDied","Data":"f155cd461d0ae85ad1062b9959124d1122212d885cb667e8395b461af6953cdd"} Mar 20 16:19:07 crc kubenswrapper[4936]: I0320 16:19:07.010902 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5233e4bc-903d-4c6a-8319-085a365268d0","Type":"ContainerDied","Data":"634b46ddadc9a4c5e4f2c417a943beef8f06198aea131fb1dbec29ce4b1ca4d9"} Mar 20 16:19:08 crc kubenswrapper[4936]: I0320 16:19:08.033817 4936 generic.go:334] "Generic (PLEG): container finished" podID="62abfd11-7276-46a9-b53a-9e90fc8ebbde" containerID="500981b2dbc9c7c4e6e72392c7242fa937724f289025c56305f488d06c177e0c" exitCode=0 Mar 20 16:19:08 crc kubenswrapper[4936]: I0320 16:19:08.034035 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qmlw5" event={"ID":"62abfd11-7276-46a9-b53a-9e90fc8ebbde","Type":"ContainerDied","Data":"500981b2dbc9c7c4e6e72392c7242fa937724f289025c56305f488d06c177e0c"} Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.463262 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74b7b699cf-b6cr9"] Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.500933 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79458c9dc6-vlzgk"] Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.504350 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.515221 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.528237 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79458c9dc6-vlzgk"] Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.539994 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-secret-key\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.540038 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-config-data\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.540055 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk2p7\" (UniqueName: \"kubernetes.io/projected/02669555-c7e4-461a-86bd-aa30db944ab9-kube-api-access-hk2p7\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.540079 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-scripts\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.540104 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02669555-c7e4-461a-86bd-aa30db944ab9-logs\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.540135 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-tls-certs\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.540166 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-combined-ca-bundle\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.571256 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-754c5bd75c-9hhl6"] Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.591044 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6ff4f57d48-qp24t"] Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.592675 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.597926 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ff4f57d48-qp24t"] Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641358 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02669555-c7e4-461a-86bd-aa30db944ab9-logs\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641409 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knb4m\" (UniqueName: \"kubernetes.io/projected/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-kube-api-access-knb4m\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641443 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-horizon-secret-key\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641471 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-tls-certs\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641492 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-combined-ca-bundle\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641512 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-horizon-tls-certs\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641559 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-combined-ca-bundle\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641596 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-config-data\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641614 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-scripts\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641639 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-logs\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641663 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-secret-key\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641698 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-config-data\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641715 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk2p7\" (UniqueName: \"kubernetes.io/projected/02669555-c7e4-461a-86bd-aa30db944ab9-kube-api-access-hk2p7\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.641793 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-scripts\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.642618 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-scripts\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.642823 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02669555-c7e4-461a-86bd-aa30db944ab9-logs\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.648895 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-secret-key\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.649635 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-combined-ca-bundle\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.653359 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-config-data\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.662207 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-tls-certs\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.682221 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk2p7\" (UniqueName: \"kubernetes.io/projected/02669555-c7e4-461a-86bd-aa30db944ab9-kube-api-access-hk2p7\") pod \"horizon-79458c9dc6-vlzgk\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.742664 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knb4m\" (UniqueName: \"kubernetes.io/projected/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-kube-api-access-knb4m\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.742717 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-horizon-secret-key\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.742756 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-combined-ca-bundle\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.742780 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-horizon-tls-certs\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.742837 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-config-data\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.742862 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-scripts\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.742893 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-logs\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.746571 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-config-data\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.746672 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-horizon-secret-key\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.747563 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-horizon-tls-certs\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.748925 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-scripts\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.749514 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-logs\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.750046 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-combined-ca-bundle\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.763846 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knb4m\" (UniqueName: \"kubernetes.io/projected/7f1bd9ee-e0b7-4ea9-b3ca-97987941f149-kube-api-access-knb4m\") pod \"horizon-6ff4f57d48-qp24t\" (UID: \"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149\") " pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.849022 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:09 crc kubenswrapper[4936]: I0320 16:19:09.916398 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:11 crc kubenswrapper[4936]: I0320 16:19:11.007732 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:19:11 crc kubenswrapper[4936]: I0320 16:19:11.077531 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-6g4tb"] Mar 20 16:19:11 crc kubenswrapper[4936]: I0320 16:19:11.077911 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="dnsmasq-dns" containerID="cri-o://8e03a6e951063b53256be6fad58cf93a73368d0fa4111e4c8bff3041a956b205" gracePeriod=10 Mar 20 16:19:11 crc kubenswrapper[4936]: I0320 16:19:11.968623 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Mar 20 16:19:12 crc kubenswrapper[4936]: I0320 16:19:12.086483 4936 generic.go:334] "Generic (PLEG): container finished" podID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerID="8e03a6e951063b53256be6fad58cf93a73368d0fa4111e4c8bff3041a956b205" exitCode=0 Mar 20 16:19:12 crc kubenswrapper[4936]: I0320 16:19:12.086538 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" event={"ID":"f53878d4-2d6f-45c0-a7d8-a143144d5061","Type":"ContainerDied","Data":"8e03a6e951063b53256be6fad58cf93a73368d0fa4111e4c8bff3041a956b205"} Mar 20 16:19:16 crc kubenswrapper[4936]: I0320 16:19:16.968900 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Mar 20 16:19:19 crc kubenswrapper[4936]: E0320 16:19:19.485494 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 20 16:19:19 crc kubenswrapper[4936]: E0320 16:19:19.485916 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n687h5c4h5fch685h678h9bhcdh8ch645hcdh58ch9h5dh76h576hfbh699h58h6bh678h68dh6fh66bh584hd8h554h85h576h5fbh567h549h66bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d2h52,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-754c5bd75c-9hhl6_openstack(19df092c-5039-4069-9010-16248fb51d14): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:19:19 crc kubenswrapper[4936]: E0320 16:19:19.498100 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-754c5bd75c-9hhl6" podUID="19df092c-5039-4069-9010-16248fb51d14" Mar 20 16:19:21 crc kubenswrapper[4936]: I0320 16:19:21.969117 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Mar 20 16:19:21 crc kubenswrapper[4936]: I0320 16:19:21.969267 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.916719 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.957119 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pxf4\" (UniqueName: \"kubernetes.io/projected/62abfd11-7276-46a9-b53a-9e90fc8ebbde-kube-api-access-6pxf4\") pod \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.957186 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-fernet-keys\") pod \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.957585 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-config-data\") pod \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.957698 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-combined-ca-bundle\") pod \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.957830 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-credential-keys\") pod \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.958106 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-scripts\") pod \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.985068 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62abfd11-7276-46a9-b53a-9e90fc8ebbde-kube-api-access-6pxf4" (OuterVolumeSpecName: "kube-api-access-6pxf4") pod "62abfd11-7276-46a9-b53a-9e90fc8ebbde" (UID: "62abfd11-7276-46a9-b53a-9e90fc8ebbde"). InnerVolumeSpecName "kube-api-access-6pxf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.989675 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-scripts" (OuterVolumeSpecName: "scripts") pod "62abfd11-7276-46a9-b53a-9e90fc8ebbde" (UID: "62abfd11-7276-46a9-b53a-9e90fc8ebbde"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:24 crc kubenswrapper[4936]: I0320 16:19:24.992114 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "62abfd11-7276-46a9-b53a-9e90fc8ebbde" (UID: "62abfd11-7276-46a9-b53a-9e90fc8ebbde"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.051457 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62abfd11-7276-46a9-b53a-9e90fc8ebbde" (UID: "62abfd11-7276-46a9-b53a-9e90fc8ebbde"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.053786 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "62abfd11-7276-46a9-b53a-9e90fc8ebbde" (UID: "62abfd11-7276-46a9-b53a-9e90fc8ebbde"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.061168 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-config-data" (OuterVolumeSpecName: "config-data") pod "62abfd11-7276-46a9-b53a-9e90fc8ebbde" (UID: "62abfd11-7276-46a9-b53a-9e90fc8ebbde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.061423 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-config-data\") pod \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\" (UID: \"62abfd11-7276-46a9-b53a-9e90fc8ebbde\") " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.061845 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.061871 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pxf4\" (UniqueName: \"kubernetes.io/projected/62abfd11-7276-46a9-b53a-9e90fc8ebbde-kube-api-access-6pxf4\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.061886 4936 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.061901 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.061913 4936 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: W0320 16:19:25.062011 4936 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/62abfd11-7276-46a9-b53a-9e90fc8ebbde/volumes/kubernetes.io~secret/config-data Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.062023 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-config-data" (OuterVolumeSpecName: "config-data") pod "62abfd11-7276-46a9-b53a-9e90fc8ebbde" (UID: "62abfd11-7276-46a9-b53a-9e90fc8ebbde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.163760 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62abfd11-7276-46a9-b53a-9e90fc8ebbde-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.194394 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qmlw5" event={"ID":"62abfd11-7276-46a9-b53a-9e90fc8ebbde","Type":"ContainerDied","Data":"9baa07eb7b9f338df72ae7582429e707f6f75e85ed07400bf5591c93e3cbbe06"} Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.194437 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qmlw5" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.194443 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9baa07eb7b9f338df72ae7582429e707f6f75e85ed07400bf5591c93e3cbbe06" Mar 20 16:19:25 crc kubenswrapper[4936]: E0320 16:19:25.442767 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Mar 20 16:19:25 crc kubenswrapper[4936]: E0320 16:19:25.443000 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndchd8hf9h54bh5fh66bh576h5bh576h55bh5fbh9dhc8hb8hd7hc9h589h5b6h67ch565h7dhbdh646h578h64fh5b5h579h8dhd4h654h657h5b5q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wrdhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(052101db-ddd5-419e-8131-fccbe44241b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.453197 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:19:25 crc kubenswrapper[4936]: E0320 16:19:25.460097 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 20 16:19:25 crc kubenswrapper[4936]: E0320 16:19:25.460292 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n545h675h94h58bh5c5h686h5bh667h59dhd6h5d9h59h68dh5c6h586h5cdhcchddh645h656hd4h8chdh694h6h55ch57h5cbh686h679h65dh8dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2kht7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-678fc86c4f-mxcbm_openstack(f4c49318-0b4d-48d2-aeb1-49e74d760e82): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:19:25 crc kubenswrapper[4936]: E0320 16:19:25.462696 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-678fc86c4f-mxcbm" podUID="f4c49318-0b4d-48d2-aeb1-49e74d760e82" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.470485 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-httpd-run\") pod \"468d0fa0-8842-4ff6-a36e-d62ed287850a\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.470567 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-logs\") pod \"468d0fa0-8842-4ff6-a36e-d62ed287850a\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.470624 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-scripts\") pod \"468d0fa0-8842-4ff6-a36e-d62ed287850a\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.470762 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkspl\" (UniqueName: \"kubernetes.io/projected/468d0fa0-8842-4ff6-a36e-d62ed287850a-kube-api-access-xkspl\") pod \"468d0fa0-8842-4ff6-a36e-d62ed287850a\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.470843 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-public-tls-certs\") pod \"468d0fa0-8842-4ff6-a36e-d62ed287850a\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.470884 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-config-data\") pod \"468d0fa0-8842-4ff6-a36e-d62ed287850a\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.470942 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-combined-ca-bundle\") pod \"468d0fa0-8842-4ff6-a36e-d62ed287850a\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.471071 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"468d0fa0-8842-4ff6-a36e-d62ed287850a\" (UID: \"468d0fa0-8842-4ff6-a36e-d62ed287850a\") " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.471248 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "468d0fa0-8842-4ff6-a36e-d62ed287850a" (UID: "468d0fa0-8842-4ff6-a36e-d62ed287850a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.471393 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-logs" (OuterVolumeSpecName: "logs") pod "468d0fa0-8842-4ff6-a36e-d62ed287850a" (UID: "468d0fa0-8842-4ff6-a36e-d62ed287850a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.472216 4936 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.472236 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/468d0fa0-8842-4ff6-a36e-d62ed287850a-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.483740 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "468d0fa0-8842-4ff6-a36e-d62ed287850a" (UID: "468d0fa0-8842-4ff6-a36e-d62ed287850a"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.509336 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-scripts" (OuterVolumeSpecName: "scripts") pod "468d0fa0-8842-4ff6-a36e-d62ed287850a" (UID: "468d0fa0-8842-4ff6-a36e-d62ed287850a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.509762 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/468d0fa0-8842-4ff6-a36e-d62ed287850a-kube-api-access-xkspl" (OuterVolumeSpecName: "kube-api-access-xkspl") pod "468d0fa0-8842-4ff6-a36e-d62ed287850a" (UID: "468d0fa0-8842-4ff6-a36e-d62ed287850a"). InnerVolumeSpecName "kube-api-access-xkspl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.535888 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "468d0fa0-8842-4ff6-a36e-d62ed287850a" (UID: "468d0fa0-8842-4ff6-a36e-d62ed287850a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.559457 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "468d0fa0-8842-4ff6-a36e-d62ed287850a" (UID: "468d0fa0-8842-4ff6-a36e-d62ed287850a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.566085 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-config-data" (OuterVolumeSpecName: "config-data") pod "468d0fa0-8842-4ff6-a36e-d62ed287850a" (UID: "468d0fa0-8842-4ff6-a36e-d62ed287850a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.574446 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.574484 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkspl\" (UniqueName: \"kubernetes.io/projected/468d0fa0-8842-4ff6-a36e-d62ed287850a-kube-api-access-xkspl\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.574502 4936 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.574516 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.574529 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/468d0fa0-8842-4ff6-a36e-d62ed287850a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.574588 4936 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.593322 4936 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 20 16:19:25 crc kubenswrapper[4936]: I0320 16:19:25.675942 4936 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.107376 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-qmlw5"] Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.118113 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-qmlw5"] Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.194928 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-w6hwd"] Mar 20 16:19:26 crc kubenswrapper[4936]: E0320 16:19:26.195284 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerName="glance-log" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.195295 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerName="glance-log" Mar 20 16:19:26 crc kubenswrapper[4936]: E0320 16:19:26.195311 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerName="glance-httpd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.195316 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerName="glance-httpd" Mar 20 16:19:26 crc kubenswrapper[4936]: E0320 16:19:26.195326 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62abfd11-7276-46a9-b53a-9e90fc8ebbde" containerName="keystone-bootstrap" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.195332 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="62abfd11-7276-46a9-b53a-9e90fc8ebbde" containerName="keystone-bootstrap" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.195570 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerName="glance-httpd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.195592 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="62abfd11-7276-46a9-b53a-9e90fc8ebbde" containerName="keystone-bootstrap" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.195600 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="468d0fa0-8842-4ff6-a36e-d62ed287850a" containerName="glance-log" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.197789 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.200860 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.201064 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.201193 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.202814 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.203041 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gbshp" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.210112 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-w6hwd"] Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.225912 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.226386 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"468d0fa0-8842-4ff6-a36e-d62ed287850a","Type":"ContainerDied","Data":"3b6887c4ee7343b4c920ba6ebeaf9b8454c0229caa0c512483792220007bcb0b"} Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.226447 4936 scope.go:117] "RemoveContainer" containerID="72abbdb093c7163b29219d6209d62fdc24d07574bb34a2715412536edf686b7a" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.287930 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.299779 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-scripts\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.299824 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psfv5\" (UniqueName: \"kubernetes.io/projected/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-kube-api-access-psfv5\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.299984 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-fernet-keys\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.300029 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-config-data\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.300063 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-combined-ca-bundle\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.300097 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-credential-keys\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.309666 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.319936 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.321715 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.327163 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.328651 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.329865 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401402 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-config-data\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401480 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-combined-ca-bundle\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401519 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-credential-keys\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401571 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slzm5\" (UniqueName: \"kubernetes.io/projected/be14f196-9b19-49e3-bc07-b642c9ae7ff8-kube-api-access-slzm5\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401608 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-scripts\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401632 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-logs\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401697 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-scripts\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401724 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psfv5\" (UniqueName: \"kubernetes.io/projected/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-kube-api-access-psfv5\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401756 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-config-data\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401806 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401846 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401928 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401967 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-fernet-keys\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.401991 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.406079 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-combined-ca-bundle\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.406701 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-credential-keys\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.416360 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-fernet-keys\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.420347 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-config-data\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.427135 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psfv5\" (UniqueName: \"kubernetes.io/projected/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-kube-api-access-psfv5\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.432038 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-scripts\") pod \"keystone-bootstrap-w6hwd\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.505090 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.505176 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.505252 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slzm5\" (UniqueName: \"kubernetes.io/projected/be14f196-9b19-49e3-bc07-b642c9ae7ff8-kube-api-access-slzm5\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.505281 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-scripts\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.505306 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-logs\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.505361 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-config-data\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.505402 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.505436 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.508294 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.525026 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-logs\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.526567 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.528436 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.536388 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.538608 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slzm5\" (UniqueName: \"kubernetes.io/projected/be14f196-9b19-49e3-bc07-b642c9ae7ff8-kube-api-access-slzm5\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.539948 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-scripts\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.552889 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.554863 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-config-data\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.599742 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.645963 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:19:26 crc kubenswrapper[4936]: I0320 16:19:26.969110 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Mar 20 16:19:27 crc kubenswrapper[4936]: E0320 16:19:27.458232 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Mar 20 16:19:27 crc kubenswrapper[4936]: E0320 16:19:27.458414 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g497k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-j4d5s_openstack(b87c4418-b13e-41e2-bbd8-d7dfb057a594): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:19:27 crc kubenswrapper[4936]: E0320 16:19:27.459848 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-j4d5s" podUID="b87c4418-b13e-41e2-bbd8-d7dfb057a594" Mar 20 16:19:27 crc kubenswrapper[4936]: E0320 16:19:27.474064 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 20 16:19:27 crc kubenswrapper[4936]: E0320 16:19:27.474479 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfh84h65dh5f9h674h6bh5fh8bh589h58fh54bh67fh59ch5fdh554h98h5d9h88hb8hf4hd5hcfh77hc9h546h5bch5d4h65fh654h649h54ch59dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jd9wf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-74b7b699cf-b6cr9_openstack(b7f8b999-b632-4354-bc26-925b418b379e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:19:27 crc kubenswrapper[4936]: E0320 16:19:27.479556 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-74b7b699cf-b6cr9" podUID="b7f8b999-b632-4354-bc26-925b418b379e" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.533072 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.623754 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-logs\") pod \"5233e4bc-903d-4c6a-8319-085a365268d0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.623826 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"5233e4bc-903d-4c6a-8319-085a365268d0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.623889 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-internal-tls-certs\") pod \"5233e4bc-903d-4c6a-8319-085a365268d0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.624023 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxxzc\" (UniqueName: \"kubernetes.io/projected/5233e4bc-903d-4c6a-8319-085a365268d0-kube-api-access-pxxzc\") pod \"5233e4bc-903d-4c6a-8319-085a365268d0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.624096 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-scripts\") pod \"5233e4bc-903d-4c6a-8319-085a365268d0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.624154 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-combined-ca-bundle\") pod \"5233e4bc-903d-4c6a-8319-085a365268d0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.624193 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-httpd-run\") pod \"5233e4bc-903d-4c6a-8319-085a365268d0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.624320 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-config-data\") pod \"5233e4bc-903d-4c6a-8319-085a365268d0\" (UID: \"5233e4bc-903d-4c6a-8319-085a365268d0\") " Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.625054 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-logs" (OuterVolumeSpecName: "logs") pod "5233e4bc-903d-4c6a-8319-085a365268d0" (UID: "5233e4bc-903d-4c6a-8319-085a365268d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.625061 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5233e4bc-903d-4c6a-8319-085a365268d0" (UID: "5233e4bc-903d-4c6a-8319-085a365268d0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.630405 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5233e4bc-903d-4c6a-8319-085a365268d0-kube-api-access-pxxzc" (OuterVolumeSpecName: "kube-api-access-pxxzc") pod "5233e4bc-903d-4c6a-8319-085a365268d0" (UID: "5233e4bc-903d-4c6a-8319-085a365268d0"). InnerVolumeSpecName "kube-api-access-pxxzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.635177 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-scripts" (OuterVolumeSpecName: "scripts") pod "5233e4bc-903d-4c6a-8319-085a365268d0" (UID: "5233e4bc-903d-4c6a-8319-085a365268d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.641110 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "5233e4bc-903d-4c6a-8319-085a365268d0" (UID: "5233e4bc-903d-4c6a-8319-085a365268d0"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.656174 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5233e4bc-903d-4c6a-8319-085a365268d0" (UID: "5233e4bc-903d-4c6a-8319-085a365268d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.685216 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-config-data" (OuterVolumeSpecName: "config-data") pod "5233e4bc-903d-4c6a-8319-085a365268d0" (UID: "5233e4bc-903d-4c6a-8319-085a365268d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.706098 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5233e4bc-903d-4c6a-8319-085a365268d0" (UID: "5233e4bc-903d-4c6a-8319-085a365268d0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.727007 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxxzc\" (UniqueName: \"kubernetes.io/projected/5233e4bc-903d-4c6a-8319-085a365268d0-kube-api-access-pxxzc\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.727061 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.727078 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.727090 4936 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.727104 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.727116 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5233e4bc-903d-4c6a-8319-085a365268d0-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.727151 4936 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.727165 4936 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5233e4bc-903d-4c6a-8319-085a365268d0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.747323 4936 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.828660 4936 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.866172 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="468d0fa0-8842-4ff6-a36e-d62ed287850a" path="/var/lib/kubelet/pods/468d0fa0-8842-4ff6-a36e-d62ed287850a/volumes" Mar 20 16:19:27 crc kubenswrapper[4936]: I0320 16:19:27.867917 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62abfd11-7276-46a9-b53a-9e90fc8ebbde" path="/var/lib/kubelet/pods/62abfd11-7276-46a9-b53a-9e90fc8ebbde/volumes" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.245149 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5233e4bc-903d-4c6a-8319-085a365268d0","Type":"ContainerDied","Data":"56922ff6855d30eef029b7c5b5d9d04f0a821d2e785094f2716d6c3a6c8ae8ff"} Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.245266 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: E0320 16:19:28.250050 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-j4d5s" podUID="b87c4418-b13e-41e2-bbd8-d7dfb057a594" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.297525 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.312943 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.320566 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:19:28 crc kubenswrapper[4936]: E0320 16:19:28.320990 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5233e4bc-903d-4c6a-8319-085a365268d0" containerName="glance-log" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.321009 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="5233e4bc-903d-4c6a-8319-085a365268d0" containerName="glance-log" Mar 20 16:19:28 crc kubenswrapper[4936]: E0320 16:19:28.321030 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5233e4bc-903d-4c6a-8319-085a365268d0" containerName="glance-httpd" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.321036 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="5233e4bc-903d-4c6a-8319-085a365268d0" containerName="glance-httpd" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.321237 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="5233e4bc-903d-4c6a-8319-085a365268d0" containerName="glance-log" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.321251 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="5233e4bc-903d-4c6a-8319-085a365268d0" containerName="glance-httpd" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.322134 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.337301 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.337514 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.351359 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.451781 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.451890 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.451914 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.451976 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.452013 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb8bs\" (UniqueName: \"kubernetes.io/projected/601c14d9-f6a4-4f7c-b033-7a35d41594a2-kube-api-access-pb8bs\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.452042 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.452082 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.452259 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.555455 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.555512 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.555579 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.555610 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.555657 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.555676 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.555708 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.555727 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb8bs\" (UniqueName: \"kubernetes.io/projected/601c14d9-f6a4-4f7c-b033-7a35d41594a2-kube-api-access-pb8bs\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.556150 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.557068 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.559684 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.579005 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.591150 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb8bs\" (UniqueName: \"kubernetes.io/projected/601c14d9-f6a4-4f7c-b033-7a35d41594a2-kube-api-access-pb8bs\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.591925 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.595350 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.605357 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.623575 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.667159 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.856569 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:19:28 crc kubenswrapper[4936]: I0320 16:19:28.856643 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:19:29 crc kubenswrapper[4936]: I0320 16:19:29.864533 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5233e4bc-903d-4c6a-8319-085a365268d0" path="/var/lib/kubelet/pods/5233e4bc-903d-4c6a-8319-085a365268d0/volumes" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.135965 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.140639 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.150935 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.157818 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298343 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-config-data\") pod \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298402 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-swift-storage-0\") pod \"f53878d4-2d6f-45c0-a7d8-a143144d5061\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298433 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-scripts\") pod \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298469 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b7f8b999-b632-4354-bc26-925b418b379e-horizon-secret-key\") pod \"b7f8b999-b632-4354-bc26-925b418b379e\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298526 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7f8b999-b632-4354-bc26-925b418b379e-logs\") pod \"b7f8b999-b632-4354-bc26-925b418b379e\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298583 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-476wm\" (UniqueName: \"kubernetes.io/projected/f53878d4-2d6f-45c0-a7d8-a143144d5061-kube-api-access-476wm\") pod \"f53878d4-2d6f-45c0-a7d8-a143144d5061\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298617 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19df092c-5039-4069-9010-16248fb51d14-logs\") pod \"19df092c-5039-4069-9010-16248fb51d14\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298651 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2h52\" (UniqueName: \"kubernetes.io/projected/19df092c-5039-4069-9010-16248fb51d14-kube-api-access-d2h52\") pod \"19df092c-5039-4069-9010-16248fb51d14\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298681 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/19df092c-5039-4069-9010-16248fb51d14-horizon-secret-key\") pod \"19df092c-5039-4069-9010-16248fb51d14\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298717 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-config-data\") pod \"b7f8b999-b632-4354-bc26-925b418b379e\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298758 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-config\") pod \"f53878d4-2d6f-45c0-a7d8-a143144d5061\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298819 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-scripts\") pod \"b7f8b999-b632-4354-bc26-925b418b379e\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298890 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kht7\" (UniqueName: \"kubernetes.io/projected/f4c49318-0b4d-48d2-aeb1-49e74d760e82-kube-api-access-2kht7\") pod \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298917 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd9wf\" (UniqueName: \"kubernetes.io/projected/b7f8b999-b632-4354-bc26-925b418b379e-kube-api-access-jd9wf\") pod \"b7f8b999-b632-4354-bc26-925b418b379e\" (UID: \"b7f8b999-b632-4354-bc26-925b418b379e\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298955 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-sb\") pod \"f53878d4-2d6f-45c0-a7d8-a143144d5061\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298959 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-scripts" (OuterVolumeSpecName: "scripts") pod "f4c49318-0b4d-48d2-aeb1-49e74d760e82" (UID: "f4c49318-0b4d-48d2-aeb1-49e74d760e82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.298999 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-config-data\") pod \"19df092c-5039-4069-9010-16248fb51d14\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.299063 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4c49318-0b4d-48d2-aeb1-49e74d760e82-logs\") pod \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.299096 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-config-data" (OuterVolumeSpecName: "config-data") pod "f4c49318-0b4d-48d2-aeb1-49e74d760e82" (UID: "f4c49318-0b4d-48d2-aeb1-49e74d760e82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.299105 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4c49318-0b4d-48d2-aeb1-49e74d760e82-horizon-secret-key\") pod \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\" (UID: \"f4c49318-0b4d-48d2-aeb1-49e74d760e82\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.299576 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-nb\") pod \"f53878d4-2d6f-45c0-a7d8-a143144d5061\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.299627 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-scripts\") pod \"19df092c-5039-4069-9010-16248fb51d14\" (UID: \"19df092c-5039-4069-9010-16248fb51d14\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.299639 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7f8b999-b632-4354-bc26-925b418b379e-logs" (OuterVolumeSpecName: "logs") pod "b7f8b999-b632-4354-bc26-925b418b379e" (UID: "b7f8b999-b632-4354-bc26-925b418b379e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.299659 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-svc\") pod \"f53878d4-2d6f-45c0-a7d8-a143144d5061\" (UID: \"f53878d4-2d6f-45c0-a7d8-a143144d5061\") " Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.299842 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-config-data" (OuterVolumeSpecName: "config-data") pod "b7f8b999-b632-4354-bc26-925b418b379e" (UID: "b7f8b999-b632-4354-bc26-925b418b379e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.300120 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4c49318-0b4d-48d2-aeb1-49e74d760e82-logs" (OuterVolumeSpecName: "logs") pod "f4c49318-0b4d-48d2-aeb1-49e74d760e82" (UID: "f4c49318-0b4d-48d2-aeb1-49e74d760e82"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.300134 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19df092c-5039-4069-9010-16248fb51d14-logs" (OuterVolumeSpecName: "logs") pod "19df092c-5039-4069-9010-16248fb51d14" (UID: "19df092c-5039-4069-9010-16248fb51d14"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.300406 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-config-data" (OuterVolumeSpecName: "config-data") pod "19df092c-5039-4069-9010-16248fb51d14" (UID: "19df092c-5039-4069-9010-16248fb51d14"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.300662 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-scripts" (OuterVolumeSpecName: "scripts") pod "b7f8b999-b632-4354-bc26-925b418b379e" (UID: "b7f8b999-b632-4354-bc26-925b418b379e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.301378 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.301413 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.301427 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4c49318-0b4d-48d2-aeb1-49e74d760e82-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.301441 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.301454 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c49318-0b4d-48d2-aeb1-49e74d760e82-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.301466 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7f8b999-b632-4354-bc26-925b418b379e-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.301477 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19df092c-5039-4069-9010-16248fb51d14-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.301488 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f8b999-b632-4354-bc26-925b418b379e-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.304081 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-scripts" (OuterVolumeSpecName: "scripts") pod "19df092c-5039-4069-9010-16248fb51d14" (UID: "19df092c-5039-4069-9010-16248fb51d14"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.305493 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19df092c-5039-4069-9010-16248fb51d14-kube-api-access-d2h52" (OuterVolumeSpecName: "kube-api-access-d2h52") pod "19df092c-5039-4069-9010-16248fb51d14" (UID: "19df092c-5039-4069-9010-16248fb51d14"). InnerVolumeSpecName "kube-api-access-d2h52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.306370 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4c49318-0b4d-48d2-aeb1-49e74d760e82-kube-api-access-2kht7" (OuterVolumeSpecName: "kube-api-access-2kht7") pod "f4c49318-0b4d-48d2-aeb1-49e74d760e82" (UID: "f4c49318-0b4d-48d2-aeb1-49e74d760e82"). InnerVolumeSpecName "kube-api-access-2kht7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.306996 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19df092c-5039-4069-9010-16248fb51d14-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "19df092c-5039-4069-9010-16248fb51d14" (UID: "19df092c-5039-4069-9010-16248fb51d14"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.308006 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4c49318-0b4d-48d2-aeb1-49e74d760e82-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f4c49318-0b4d-48d2-aeb1-49e74d760e82" (UID: "f4c49318-0b4d-48d2-aeb1-49e74d760e82"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.308014 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f53878d4-2d6f-45c0-a7d8-a143144d5061-kube-api-access-476wm" (OuterVolumeSpecName: "kube-api-access-476wm") pod "f53878d4-2d6f-45c0-a7d8-a143144d5061" (UID: "f53878d4-2d6f-45c0-a7d8-a143144d5061"). InnerVolumeSpecName "kube-api-access-476wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.308911 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7f8b999-b632-4354-bc26-925b418b379e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b7f8b999-b632-4354-bc26-925b418b379e" (UID: "b7f8b999-b632-4354-bc26-925b418b379e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.314472 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f8b999-b632-4354-bc26-925b418b379e-kube-api-access-jd9wf" (OuterVolumeSpecName: "kube-api-access-jd9wf") pod "b7f8b999-b632-4354-bc26-925b418b379e" (UID: "b7f8b999-b632-4354-bc26-925b418b379e"). InnerVolumeSpecName "kube-api-access-jd9wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.321916 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678fc86c4f-mxcbm" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.323397 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678fc86c4f-mxcbm" event={"ID":"f4c49318-0b4d-48d2-aeb1-49e74d760e82","Type":"ContainerDied","Data":"692ca8f29e2d6c576b1814ba84ef663ffafb77475156dc143c93fd6e335067c9"} Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.331000 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b7b699cf-b6cr9" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.331071 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b7b699cf-b6cr9" event={"ID":"b7f8b999-b632-4354-bc26-925b418b379e","Type":"ContainerDied","Data":"deedeb6f023a16f466a478a4a1010b30345eb3d2376f90b037f0879ceb59d814"} Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.335281 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754c5bd75c-9hhl6" event={"ID":"19df092c-5039-4069-9010-16248fb51d14","Type":"ContainerDied","Data":"6c88fa74ccbe8c6e16335e3c79b0c3e6ca9792b04e3ae2fb25fdb773eb4007bb"} Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.335316 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754c5bd75c-9hhl6" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.339918 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" event={"ID":"f53878d4-2d6f-45c0-a7d8-a143144d5061","Type":"ContainerDied","Data":"8ff681eddc0a3906d98c4446488b70a2cdec3e5968cb43dd86e6c3bc3547768e"} Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.340039 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.355239 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f53878d4-2d6f-45c0-a7d8-a143144d5061" (UID: "f53878d4-2d6f-45c0-a7d8-a143144d5061"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.358804 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f53878d4-2d6f-45c0-a7d8-a143144d5061" (UID: "f53878d4-2d6f-45c0-a7d8-a143144d5061"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.399995 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f53878d4-2d6f-45c0-a7d8-a143144d5061" (UID: "f53878d4-2d6f-45c0-a7d8-a143144d5061"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.400340 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f53878d4-2d6f-45c0-a7d8-a143144d5061" (UID: "f53878d4-2d6f-45c0-a7d8-a143144d5061"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402700 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-476wm\" (UniqueName: \"kubernetes.io/projected/f53878d4-2d6f-45c0-a7d8-a143144d5061-kube-api-access-476wm\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402742 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2h52\" (UniqueName: \"kubernetes.io/projected/19df092c-5039-4069-9010-16248fb51d14-kube-api-access-d2h52\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402755 4936 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/19df092c-5039-4069-9010-16248fb51d14-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402766 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kht7\" (UniqueName: \"kubernetes.io/projected/f4c49318-0b4d-48d2-aeb1-49e74d760e82-kube-api-access-2kht7\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402775 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd9wf\" (UniqueName: \"kubernetes.io/projected/b7f8b999-b632-4354-bc26-925b418b379e-kube-api-access-jd9wf\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402784 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402794 4936 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4c49318-0b4d-48d2-aeb1-49e74d760e82-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402801 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402809 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19df092c-5039-4069-9010-16248fb51d14-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402816 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402824 4936 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.402834 4936 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b7f8b999-b632-4354-bc26-925b418b379e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.413397 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-config" (OuterVolumeSpecName: "config") pod "f53878d4-2d6f-45c0-a7d8-a143144d5061" (UID: "f53878d4-2d6f-45c0-a7d8-a143144d5061"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.505422 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f53878d4-2d6f-45c0-a7d8-a143144d5061-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.526992 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-754c5bd75c-9hhl6"] Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.537199 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-754c5bd75c-9hhl6"] Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.570298 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74b7b699cf-b6cr9"] Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.585094 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-74b7b699cf-b6cr9"] Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.605764 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-678fc86c4f-mxcbm"] Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.614432 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-678fc86c4f-mxcbm"] Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.679917 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-6g4tb"] Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.689434 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-6g4tb"] Mar 20 16:19:36 crc kubenswrapper[4936]: I0320 16:19:36.969249 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-6g4tb" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Mar 20 16:19:37 crc kubenswrapper[4936]: I0320 16:19:37.864903 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19df092c-5039-4069-9010-16248fb51d14" path="/var/lib/kubelet/pods/19df092c-5039-4069-9010-16248fb51d14/volumes" Mar 20 16:19:37 crc kubenswrapper[4936]: I0320 16:19:37.865368 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7f8b999-b632-4354-bc26-925b418b379e" path="/var/lib/kubelet/pods/b7f8b999-b632-4354-bc26-925b418b379e/volumes" Mar 20 16:19:37 crc kubenswrapper[4936]: I0320 16:19:37.865852 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4c49318-0b4d-48d2-aeb1-49e74d760e82" path="/var/lib/kubelet/pods/f4c49318-0b4d-48d2-aeb1-49e74d760e82/volumes" Mar 20 16:19:37 crc kubenswrapper[4936]: I0320 16:19:37.866281 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" path="/var/lib/kubelet/pods/f53878d4-2d6f-45c0-a7d8-a143144d5061/volumes" Mar 20 16:19:38 crc kubenswrapper[4936]: I0320 16:19:38.341263 4936 scope.go:117] "RemoveContainer" containerID="0f700379c382adf0d0212cfd0a1f3e8211bc897cbefd37061acee29757acc695" Mar 20 16:19:38 crc kubenswrapper[4936]: E0320 16:19:38.370611 4936 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 20 16:19:38 crc kubenswrapper[4936]: E0320 16:19:38.370829 4936 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j49gd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-n7kq7_openstack(047fb63e-64a8-4280-9846-659575038df3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 16:19:38 crc kubenswrapper[4936]: E0320 16:19:38.372086 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-n7kq7" podUID="047fb63e-64a8-4280-9846-659575038df3" Mar 20 16:19:38 crc kubenswrapper[4936]: I0320 16:19:38.812806 4936 scope.go:117] "RemoveContainer" containerID="f155cd461d0ae85ad1062b9959124d1122212d885cb667e8395b461af6953cdd" Mar 20 16:19:38 crc kubenswrapper[4936]: I0320 16:19:38.918833 4936 scope.go:117] "RemoveContainer" containerID="634b46ddadc9a4c5e4f2c417a943beef8f06198aea131fb1dbec29ce4b1ca4d9" Mar 20 16:19:38 crc kubenswrapper[4936]: I0320 16:19:38.995735 4936 scope.go:117] "RemoveContainer" containerID="8e03a6e951063b53256be6fad58cf93a73368d0fa4111e4c8bff3041a956b205" Mar 20 16:19:39 crc kubenswrapper[4936]: I0320 16:19:39.052145 4936 scope.go:117] "RemoveContainer" containerID="2d8acde0827e044d41df774e7e14759e88630ce148f1aca6d31bcfbc0fe8c3c9" Mar 20 16:19:39 crc kubenswrapper[4936]: I0320 16:19:39.376419 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-w6hwd"] Mar 20 16:19:39 crc kubenswrapper[4936]: I0320 16:19:39.385297 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9w2gg" event={"ID":"61aff63e-dc1b-4854-af4a-188837ad4cf9","Type":"ContainerStarted","Data":"5c9803b6aad8ae456f72f069496c28ed3be945dee4fe018b3c421ef016932f5e"} Mar 20 16:19:39 crc kubenswrapper[4936]: E0320 16:19:39.389738 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-n7kq7" podUID="047fb63e-64a8-4280-9846-659575038df3" Mar 20 16:19:39 crc kubenswrapper[4936]: I0320 16:19:39.421346 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9w2gg" podStartSLOduration=5.594182416 podStartE2EDuration="39.421321095s" podCreationTimestamp="2026-03-20 16:19:00 +0000 UTC" firstStartedPulling="2026-03-20 16:19:02.220618749 +0000 UTC m=+1093.166986564" lastFinishedPulling="2026-03-20 16:19:36.047757438 +0000 UTC m=+1126.994125243" observedRunningTime="2026-03-20 16:19:39.404945897 +0000 UTC m=+1130.351313722" watchObservedRunningTime="2026-03-20 16:19:39.421321095 +0000 UTC m=+1130.367688910" Mar 20 16:19:39 crc kubenswrapper[4936]: W0320 16:19:39.492137 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f1bd9ee_e0b7_4ea9_b3ca_97987941f149.slice/crio-626b8eb9fb9731ebff03316e314d1ffff4941eaad88a02190591823dc9e29e53 WatchSource:0}: Error finding container 626b8eb9fb9731ebff03316e314d1ffff4941eaad88a02190591823dc9e29e53: Status 404 returned error can't find the container with id 626b8eb9fb9731ebff03316e314d1ffff4941eaad88a02190591823dc9e29e53 Mar 20 16:19:39 crc kubenswrapper[4936]: I0320 16:19:39.498684 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79458c9dc6-vlzgk"] Mar 20 16:19:39 crc kubenswrapper[4936]: I0320 16:19:39.516892 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ff4f57d48-qp24t"] Mar 20 16:19:39 crc kubenswrapper[4936]: I0320 16:19:39.559280 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.413429 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff4f57d48-qp24t" event={"ID":"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149","Type":"ContainerStarted","Data":"626b8eb9fb9731ebff03316e314d1ffff4941eaad88a02190591823dc9e29e53"} Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.433922 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.434425 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79458c9dc6-vlzgk" event={"ID":"02669555-c7e4-461a-86bd-aa30db944ab9","Type":"ContainerStarted","Data":"ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4"} Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.434464 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79458c9dc6-vlzgk" event={"ID":"02669555-c7e4-461a-86bd-aa30db944ab9","Type":"ContainerStarted","Data":"37f336d8f70e8fbe50557df41e8f7a1992229c8786b5c84e5558dc0a7a22f068"} Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.440340 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be14f196-9b19-49e3-bc07-b642c9ae7ff8","Type":"ContainerStarted","Data":"9433d1ec221bb76adc3c454f903b2634b9e3a2292e74d1bbd6adb254492f958d"} Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.440386 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be14f196-9b19-49e3-bc07-b642c9ae7ff8","Type":"ContainerStarted","Data":"0117ecdb8c9b0d9c85684fa1f18277bfb3211cc39c579b96b4ea9282d9bf8bbe"} Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.444332 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"052101db-ddd5-419e-8131-fccbe44241b4","Type":"ContainerStarted","Data":"38e37211aa92495ebb6cf8608eefe13a16c2bebaa4f1fcedc52ab95d9fa21412"} Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.447914 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-w6hwd" event={"ID":"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755","Type":"ContainerStarted","Data":"6ea8047ac48d471af2f2db2043810141564ff85eb8ce2ea84b1c0e7da951f117"} Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.447993 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-w6hwd" event={"ID":"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755","Type":"ContainerStarted","Data":"4f122fede981cda6d5d9040735c13f191181f50c7434686f8776ffdd3e5eb2d7"} Mar 20 16:19:40 crc kubenswrapper[4936]: I0320 16:19:40.467300 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-w6hwd" podStartSLOduration=14.467285259 podStartE2EDuration="14.467285259s" podCreationTimestamp="2026-03-20 16:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:40.462758247 +0000 UTC m=+1131.409126062" watchObservedRunningTime="2026-03-20 16:19:40.467285259 +0000 UTC m=+1131.413653074" Mar 20 16:19:41 crc kubenswrapper[4936]: I0320 16:19:41.480304 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be14f196-9b19-49e3-bc07-b642c9ae7ff8","Type":"ContainerStarted","Data":"1ca8e48a473928080395ecdceb6e5c0280a91455aee8c60600269c60c57486d4"} Mar 20 16:19:41 crc kubenswrapper[4936]: I0320 16:19:41.495201 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff4f57d48-qp24t" event={"ID":"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149","Type":"ContainerStarted","Data":"146b5a28aee8c175e11733fec77a6da2dbdcc086d1e53cd4f8fd6536d421102a"} Mar 20 16:19:41 crc kubenswrapper[4936]: I0320 16:19:41.495251 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff4f57d48-qp24t" event={"ID":"7f1bd9ee-e0b7-4ea9-b3ca-97987941f149","Type":"ContainerStarted","Data":"d52d93232f9732032f7038ebf1b1529e9fa122505da013eca220725f3ee04cdf"} Mar 20 16:19:41 crc kubenswrapper[4936]: I0320 16:19:41.543011 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=15.542994458999999 podStartE2EDuration="15.542994459s" podCreationTimestamp="2026-03-20 16:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:41.5419932 +0000 UTC m=+1132.488361015" watchObservedRunningTime="2026-03-20 16:19:41.542994459 +0000 UTC m=+1132.489362274" Mar 20 16:19:41 crc kubenswrapper[4936]: I0320 16:19:41.543527 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79458c9dc6-vlzgk" event={"ID":"02669555-c7e4-461a-86bd-aa30db944ab9","Type":"ContainerStarted","Data":"b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9"} Mar 20 16:19:41 crc kubenswrapper[4936]: I0320 16:19:41.570573 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"601c14d9-f6a4-4f7c-b033-7a35d41594a2","Type":"ContainerStarted","Data":"5d07190ffa6702b331b01eafb1e8f1975d312c0eb6d60b7f88fce924ab3951af"} Mar 20 16:19:41 crc kubenswrapper[4936]: I0320 16:19:41.570623 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"601c14d9-f6a4-4f7c-b033-7a35d41594a2","Type":"ContainerStarted","Data":"8601f72ec3acf0b1e58ec2f8002d449a7e5235cf88ca804977b7ac869c54f6f0"} Mar 20 16:19:41 crc kubenswrapper[4936]: I0320 16:19:41.599247 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6ff4f57d48-qp24t" podStartSLOduration=31.900451578 podStartE2EDuration="32.5992258s" podCreationTimestamp="2026-03-20 16:19:09 +0000 UTC" firstStartedPulling="2026-03-20 16:19:39.496841839 +0000 UTC m=+1130.443209654" lastFinishedPulling="2026-03-20 16:19:40.195616061 +0000 UTC m=+1131.141983876" observedRunningTime="2026-03-20 16:19:41.5978649 +0000 UTC m=+1132.544232715" watchObservedRunningTime="2026-03-20 16:19:41.5992258 +0000 UTC m=+1132.545593615" Mar 20 16:19:42 crc kubenswrapper[4936]: I0320 16:19:42.581722 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"601c14d9-f6a4-4f7c-b033-7a35d41594a2","Type":"ContainerStarted","Data":"b4d22dfeb13ddda6b999a4c43aab3870c668690858a2a94abfcfdc556416edd3"} Mar 20 16:19:42 crc kubenswrapper[4936]: I0320 16:19:42.604533 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=14.604499686 podStartE2EDuration="14.604499686s" podCreationTimestamp="2026-03-20 16:19:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:42.601919831 +0000 UTC m=+1133.548287676" watchObservedRunningTime="2026-03-20 16:19:42.604499686 +0000 UTC m=+1133.550867501" Mar 20 16:19:42 crc kubenswrapper[4936]: I0320 16:19:42.611622 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79458c9dc6-vlzgk" podStartSLOduration=32.95699121 podStartE2EDuration="33.611601133s" podCreationTimestamp="2026-03-20 16:19:09 +0000 UTC" firstStartedPulling="2026-03-20 16:19:39.478840504 +0000 UTC m=+1130.425208309" lastFinishedPulling="2026-03-20 16:19:40.133450417 +0000 UTC m=+1131.079818232" observedRunningTime="2026-03-20 16:19:41.647915031 +0000 UTC m=+1132.594282846" watchObservedRunningTime="2026-03-20 16:19:42.611601133 +0000 UTC m=+1133.557968958" Mar 20 16:19:44 crc kubenswrapper[4936]: I0320 16:19:44.607484 4936 generic.go:334] "Generic (PLEG): container finished" podID="3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" containerID="6ea8047ac48d471af2f2db2043810141564ff85eb8ce2ea84b1c0e7da951f117" exitCode=0 Mar 20 16:19:44 crc kubenswrapper[4936]: I0320 16:19:44.607577 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-w6hwd" event={"ID":"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755","Type":"ContainerDied","Data":"6ea8047ac48d471af2f2db2043810141564ff85eb8ce2ea84b1c0e7da951f117"} Mar 20 16:19:45 crc kubenswrapper[4936]: I0320 16:19:45.626797 4936 generic.go:334] "Generic (PLEG): container finished" podID="5421152c-d474-428f-9554-96e036719655" containerID="7c4fc58febfccc4dbd4458de7a1dee9f847f11da75accafe6e2b829b70cd92f0" exitCode=0 Mar 20 16:19:45 crc kubenswrapper[4936]: I0320 16:19:45.626878 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4t65w" event={"ID":"5421152c-d474-428f-9554-96e036719655","Type":"ContainerDied","Data":"7c4fc58febfccc4dbd4458de7a1dee9f847f11da75accafe6e2b829b70cd92f0"} Mar 20 16:19:46 crc kubenswrapper[4936]: I0320 16:19:46.647162 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 20 16:19:46 crc kubenswrapper[4936]: I0320 16:19:46.647517 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 20 16:19:46 crc kubenswrapper[4936]: I0320 16:19:46.684838 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 20 16:19:46 crc kubenswrapper[4936]: I0320 16:19:46.688555 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 20 16:19:47 crc kubenswrapper[4936]: I0320 16:19:47.645077 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 20 16:19:47 crc kubenswrapper[4936]: I0320 16:19:47.645399 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 20 16:19:48 crc kubenswrapper[4936]: I0320 16:19:48.668812 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:48 crc kubenswrapper[4936]: I0320 16:19:48.669266 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:48 crc kubenswrapper[4936]: I0320 16:19:48.702237 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:48 crc kubenswrapper[4936]: I0320 16:19:48.710037 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.497363 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.507347 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.601418 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-combined-ca-bundle\") pod \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.601494 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-fernet-keys\") pod \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.601578 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psfv5\" (UniqueName: \"kubernetes.io/projected/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-kube-api-access-psfv5\") pod \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.601622 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-scripts\") pod \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.601714 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-config-data\") pod \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.601796 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-credential-keys\") pod \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\" (UID: \"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755\") " Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.608796 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-kube-api-access-psfv5" (OuterVolumeSpecName: "kube-api-access-psfv5") pod "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" (UID: "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755"). InnerVolumeSpecName "kube-api-access-psfv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.608905 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-scripts" (OuterVolumeSpecName: "scripts") pod "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" (UID: "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.620314 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" (UID: "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.626495 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" (UID: "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.646932 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.647267 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-config-data" (OuterVolumeSpecName: "config-data") pod "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" (UID: "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.670629 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4t65w" event={"ID":"5421152c-d474-428f-9554-96e036719655","Type":"ContainerDied","Data":"f8f86ce6dceba12a1119e48f55635df8ffc6e779d69448863acb2ddfa767147d"} Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.670714 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8f86ce6dceba12a1119e48f55635df8ffc6e779d69448863acb2ddfa767147d" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.670808 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4t65w" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.673586 4936 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.674577 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-w6hwd" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.674746 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-w6hwd" event={"ID":"3cd7ee68-8aa9-4b2c-9d06-c4de9698b755","Type":"ContainerDied","Data":"4f122fede981cda6d5d9040735c13f191181f50c7434686f8776ffdd3e5eb2d7"} Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.674767 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f122fede981cda6d5d9040735c13f191181f50c7434686f8776ffdd3e5eb2d7" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.676674 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.676743 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.699043 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" (UID: "3cd7ee68-8aa9-4b2c-9d06-c4de9698b755"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.703171 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-combined-ca-bundle\") pod \"5421152c-d474-428f-9554-96e036719655\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.703273 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw2x5\" (UniqueName: \"kubernetes.io/projected/5421152c-d474-428f-9554-96e036719655-kube-api-access-rw2x5\") pod \"5421152c-d474-428f-9554-96e036719655\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.703326 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-config\") pod \"5421152c-d474-428f-9554-96e036719655\" (UID: \"5421152c-d474-428f-9554-96e036719655\") " Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.703758 4936 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.703772 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.703782 4936 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.703793 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psfv5\" (UniqueName: \"kubernetes.io/projected/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-kube-api-access-psfv5\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.703807 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.703817 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.715858 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5421152c-d474-428f-9554-96e036719655-kube-api-access-rw2x5" (OuterVolumeSpecName: "kube-api-access-rw2x5") pod "5421152c-d474-428f-9554-96e036719655" (UID: "5421152c-d474-428f-9554-96e036719655"). InnerVolumeSpecName "kube-api-access-rw2x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.738746 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-config" (OuterVolumeSpecName: "config") pod "5421152c-d474-428f-9554-96e036719655" (UID: "5421152c-d474-428f-9554-96e036719655"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.748772 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5421152c-d474-428f-9554-96e036719655" (UID: "5421152c-d474-428f-9554-96e036719655"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.805433 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.805474 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw2x5\" (UniqueName: \"kubernetes.io/projected/5421152c-d474-428f-9554-96e036719655-kube-api-access-rw2x5\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.805486 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5421152c-d474-428f-9554-96e036719655-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.849441 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.849492 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.922527 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:49 crc kubenswrapper[4936]: I0320 16:19:49.922608 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.071746 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.643415 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-768674d4fd-4vgnw"] Mar 20 16:19:50 crc kubenswrapper[4936]: E0320 16:19:50.644170 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" containerName="keystone-bootstrap" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.644185 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" containerName="keystone-bootstrap" Mar 20 16:19:50 crc kubenswrapper[4936]: E0320 16:19:50.644198 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5421152c-d474-428f-9554-96e036719655" containerName="neutron-db-sync" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.644205 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="5421152c-d474-428f-9554-96e036719655" containerName="neutron-db-sync" Mar 20 16:19:50 crc kubenswrapper[4936]: E0320 16:19:50.644216 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="init" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.644222 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="init" Mar 20 16:19:50 crc kubenswrapper[4936]: E0320 16:19:50.644253 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="dnsmasq-dns" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.644260 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="dnsmasq-dns" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.644422 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f53878d4-2d6f-45c0-a7d8-a143144d5061" containerName="dnsmasq-dns" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.644438 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" containerName="keystone-bootstrap" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.644457 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="5421152c-d474-428f-9554-96e036719655" containerName="neutron-db-sync" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.656384 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.659494 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.660444 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gbshp" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.661198 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.661364 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.662816 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.663032 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.669218 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-768674d4fd-4vgnw"] Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.803088 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-s95m6"] Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.817959 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.819420 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-s95m6"] Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.826618 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prtfb\" (UniqueName: \"kubernetes.io/projected/9c24263c-e99e-4126-9428-55ea6d8cc908-kube-api-access-prtfb\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.826680 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-fernet-keys\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.826747 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-internal-tls-certs\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.826767 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-public-tls-certs\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.826854 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-credential-keys\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.826884 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-scripts\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.826941 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-combined-ca-bundle\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.826969 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-config-data\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.928738 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.928787 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.928824 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-config\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.928857 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-credential-keys\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.928875 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.928915 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-scripts\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.928968 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x5lp\" (UniqueName: \"kubernetes.io/projected/70c0f7b6-fa50-4b4a-9291-596d0b403576-kube-api-access-4x5lp\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.928989 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-combined-ca-bundle\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.929026 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-config-data\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.929072 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-svc\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.929113 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prtfb\" (UniqueName: \"kubernetes.io/projected/9c24263c-e99e-4126-9428-55ea6d8cc908-kube-api-access-prtfb\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.929187 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-fernet-keys\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.929213 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-internal-tls-certs\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.929240 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-public-tls-certs\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.934047 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-public-tls-certs\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.934787 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-scripts\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.938442 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-credential-keys\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.939514 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-fernet-keys\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.939890 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-internal-tls-certs\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.940513 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-config-data\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.940916 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c24263c-e99e-4126-9428-55ea6d8cc908-combined-ca-bundle\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:50 crc kubenswrapper[4936]: I0320 16:19:50.954597 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prtfb\" (UniqueName: \"kubernetes.io/projected/9c24263c-e99e-4126-9428-55ea6d8cc908-kube-api-access-prtfb\") pod \"keystone-768674d4fd-4vgnw\" (UID: \"9c24263c-e99e-4126-9428-55ea6d8cc908\") " pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.030496 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.030576 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.030615 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-config\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.030646 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.030713 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x5lp\" (UniqueName: \"kubernetes.io/projected/70c0f7b6-fa50-4b4a-9291-596d0b403576-kube-api-access-4x5lp\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.030763 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-svc\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.031859 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-svc\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.032468 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.033083 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.034222 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.034456 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-config\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.042353 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.059683 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c66f596d-qcxkb"] Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.061335 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.069277 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.069643 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.069889 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.071209 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-txrnm" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.073753 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x5lp\" (UniqueName: \"kubernetes.io/projected/70c0f7b6-fa50-4b4a-9291-596d0b403576-kube-api-access-4x5lp\") pod \"dnsmasq-dns-55f844cf75-s95m6\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.095683 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c66f596d-qcxkb"] Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.135909 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-combined-ca-bundle\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.136139 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjjsh\" (UniqueName: \"kubernetes.io/projected/7da58715-94c1-487a-9f20-f008b4bea305-kube-api-access-zjjsh\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.136208 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-httpd-config\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.136271 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-config\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.136385 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-ovndb-tls-certs\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.165460 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.238214 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjjsh\" (UniqueName: \"kubernetes.io/projected/7da58715-94c1-487a-9f20-f008b4bea305-kube-api-access-zjjsh\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.238281 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-httpd-config\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.238322 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-config\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.238352 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-ovndb-tls-certs\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.238387 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-combined-ca-bundle\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.245331 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-combined-ca-bundle\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.249121 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-httpd-config\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.252607 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-config\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.259062 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-ovndb-tls-certs\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.274717 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjjsh\" (UniqueName: \"kubernetes.io/projected/7da58715-94c1-487a-9f20-f008b4bea305-kube-api-access-zjjsh\") pod \"neutron-c66f596d-qcxkb\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.445799 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.559003 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-768674d4fd-4vgnw"] Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.786328 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j4d5s" event={"ID":"b87c4418-b13e-41e2-bbd8-d7dfb057a594","Type":"ContainerStarted","Data":"198ef6b866a3a9d1b7dc2c7023f673c2fcccad5662c4896ba53b259233660520"} Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.801056 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"052101db-ddd5-419e-8131-fccbe44241b4","Type":"ContainerStarted","Data":"d672b85638441f903d415abe7efb0ea9860db5fac56d38f39f99b39e37d243f7"} Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.822743 4936 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.822776 4936 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.823625 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-768674d4fd-4vgnw" event={"ID":"9c24263c-e99e-4126-9428-55ea6d8cc908","Type":"ContainerStarted","Data":"3cc413e86acca39150d44911815848ee9cc0dbd1fcf13a7d74f0351b69efc4b9"} Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.854089 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-j4d5s" podStartSLOduration=4.147271678 podStartE2EDuration="51.854066847s" podCreationTimestamp="2026-03-20 16:19:00 +0000 UTC" firstStartedPulling="2026-03-20 16:19:02.338677249 +0000 UTC m=+1093.285045064" lastFinishedPulling="2026-03-20 16:19:50.045472418 +0000 UTC m=+1140.991840233" observedRunningTime="2026-03-20 16:19:51.838034949 +0000 UTC m=+1142.784402764" watchObservedRunningTime="2026-03-20 16:19:51.854066847 +0000 UTC m=+1142.800434662" Mar 20 16:19:51 crc kubenswrapper[4936]: I0320 16:19:51.991671 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-s95m6"] Mar 20 16:19:52 crc kubenswrapper[4936]: I0320 16:19:52.172012 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c66f596d-qcxkb"] Mar 20 16:19:52 crc kubenswrapper[4936]: I0320 16:19:52.636327 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:52 crc kubenswrapper[4936]: I0320 16:19:52.641445 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 20 16:19:52 crc kubenswrapper[4936]: I0320 16:19:52.834125 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" event={"ID":"70c0f7b6-fa50-4b4a-9291-596d0b403576","Type":"ContainerStarted","Data":"a3d333ccdb3ffa3bfd1151e1da820b55d65909d5c66aff195484d7c573c11856"} Mar 20 16:19:52 crc kubenswrapper[4936]: I0320 16:19:52.838047 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c66f596d-qcxkb" event={"ID":"7da58715-94c1-487a-9f20-f008b4bea305","Type":"ContainerStarted","Data":"bc91e8a47866e95e60f1cd4efb677f88684ce2b83cace52de98a108e74f91c3f"} Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.842277 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-698b9689fc-cph26"] Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.846925 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.856261 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.856344 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.869219 4936 generic.go:334] "Generic (PLEG): container finished" podID="70c0f7b6-fa50-4b4a-9291-596d0b403576" containerID="57126127a3e4442f23fa9adde47e5ec10b626fe50b533569e9e19edd1a95cda8" exitCode=0 Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.878115 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" event={"ID":"70c0f7b6-fa50-4b4a-9291-596d0b403576","Type":"ContainerDied","Data":"57126127a3e4442f23fa9adde47e5ec10b626fe50b533569e9e19edd1a95cda8"} Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.878149 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-768674d4fd-4vgnw" event={"ID":"9c24263c-e99e-4126-9428-55ea6d8cc908","Type":"ContainerStarted","Data":"c1a6a0fce8210bf480c49fbd65caea3564abccefea6fdfeb87f306a5fbf360ae"} Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.878164 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.878175 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-698b9689fc-cph26"] Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.914334 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c66f596d-qcxkb" event={"ID":"7da58715-94c1-487a-9f20-f008b4bea305","Type":"ContainerStarted","Data":"5708ae6c1402aa64eda9c419b36f83f475ee4a5b700085736fe1fb6898b678fa"} Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.914384 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.914400 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c66f596d-qcxkb" event={"ID":"7da58715-94c1-487a-9f20-f008b4bea305","Type":"ContainerStarted","Data":"4cd46a430f3c3a8c5a6186f91c234b96eae73ad4dbe2e65fc9287325d41fe261"} Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.929515 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-httpd-config\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.929598 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-config\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.929631 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcsj2\" (UniqueName: \"kubernetes.io/projected/7c402d19-99d9-49b2-9304-12ef96435d7c-kube-api-access-lcsj2\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.929649 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-public-tls-certs\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.929731 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-combined-ca-bundle\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.929812 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-internal-tls-certs\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.929835 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-ovndb-tls-certs\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.955620 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-768674d4fd-4vgnw" podStartSLOduration=3.955600973 podStartE2EDuration="3.955600973s" podCreationTimestamp="2026-03-20 16:19:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:53.945437436 +0000 UTC m=+1144.891805251" watchObservedRunningTime="2026-03-20 16:19:53.955600973 +0000 UTC m=+1144.901968788" Mar 20 16:19:53 crc kubenswrapper[4936]: I0320 16:19:53.988286 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c66f596d-qcxkb" podStartSLOduration=2.988263116 podStartE2EDuration="2.988263116s" podCreationTimestamp="2026-03-20 16:19:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:53.97127459 +0000 UTC m=+1144.917642405" watchObservedRunningTime="2026-03-20 16:19:53.988263116 +0000 UTC m=+1144.934630931" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.031272 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-ovndb-tls-certs\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.031343 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-httpd-config\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.031402 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-config\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.031471 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcsj2\" (UniqueName: \"kubernetes.io/projected/7c402d19-99d9-49b2-9304-12ef96435d7c-kube-api-access-lcsj2\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.031511 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-public-tls-certs\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.031959 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-combined-ca-bundle\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.032274 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-internal-tls-certs\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.047736 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-public-tls-certs\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.049677 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-httpd-config\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.050779 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-config\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.052533 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-ovndb-tls-certs\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.054393 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-internal-tls-certs\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.058469 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcsj2\" (UniqueName: \"kubernetes.io/projected/7c402d19-99d9-49b2-9304-12ef96435d7c-kube-api-access-lcsj2\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.062349 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c402d19-99d9-49b2-9304-12ef96435d7c-combined-ca-bundle\") pod \"neutron-698b9689fc-cph26\" (UID: \"7c402d19-99d9-49b2-9304-12ef96435d7c\") " pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.184230 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.954902 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" event={"ID":"70c0f7b6-fa50-4b4a-9291-596d0b403576","Type":"ContainerStarted","Data":"81d0c8cc8be8bd2d4588928bd77ac428b077b959169acc5b8bbdbdc9264380d8"} Mar 20 16:19:54 crc kubenswrapper[4936]: I0320 16:19:54.967370 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:19:55 crc kubenswrapper[4936]: I0320 16:19:55.004656 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-698b9689fc-cph26"] Mar 20 16:19:55 crc kubenswrapper[4936]: I0320 16:19:55.011689 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" podStartSLOduration=5.011667021 podStartE2EDuration="5.011667021s" podCreationTimestamp="2026-03-20 16:19:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:54.980869552 +0000 UTC m=+1145.927237367" watchObservedRunningTime="2026-03-20 16:19:55.011667021 +0000 UTC m=+1145.958034836" Mar 20 16:19:55 crc kubenswrapper[4936]: I0320 16:19:55.964036 4936 generic.go:334] "Generic (PLEG): container finished" podID="61aff63e-dc1b-4854-af4a-188837ad4cf9" containerID="5c9803b6aad8ae456f72f069496c28ed3be945dee4fe018b3c421ef016932f5e" exitCode=0 Mar 20 16:19:55 crc kubenswrapper[4936]: I0320 16:19:55.964265 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9w2gg" event={"ID":"61aff63e-dc1b-4854-af4a-188837ad4cf9","Type":"ContainerDied","Data":"5c9803b6aad8ae456f72f069496c28ed3be945dee4fe018b3c421ef016932f5e"} Mar 20 16:19:55 crc kubenswrapper[4936]: I0320 16:19:55.966575 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n7kq7" event={"ID":"047fb63e-64a8-4280-9846-659575038df3","Type":"ContainerStarted","Data":"35818d37873ba37611214468b71d71e9107f339eac76ba0c3a162bac351a8d84"} Mar 20 16:19:55 crc kubenswrapper[4936]: I0320 16:19:55.973419 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-698b9689fc-cph26" event={"ID":"7c402d19-99d9-49b2-9304-12ef96435d7c","Type":"ContainerStarted","Data":"d443ab6bc51521e172415f74005ccfdf3b4a74df28880cca20ef317b2d738409"} Mar 20 16:19:55 crc kubenswrapper[4936]: I0320 16:19:55.973455 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-698b9689fc-cph26" event={"ID":"7c402d19-99d9-49b2-9304-12ef96435d7c","Type":"ContainerStarted","Data":"4505447c448834afd952922c155eed96aa30721e2a83b5d6636ed86d98d7690b"} Mar 20 16:19:55 crc kubenswrapper[4936]: I0320 16:19:55.973467 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:19:55 crc kubenswrapper[4936]: I0320 16:19:55.973476 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-698b9689fc-cph26" event={"ID":"7c402d19-99d9-49b2-9304-12ef96435d7c","Type":"ContainerStarted","Data":"58364f19cfa2598775deaa071b205331a258ba310da3d23e2911047073efe38c"} Mar 20 16:19:56 crc kubenswrapper[4936]: I0320 16:19:56.028315 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-698b9689fc-cph26" podStartSLOduration=3.028293788 podStartE2EDuration="3.028293788s" podCreationTimestamp="2026-03-20 16:19:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:19:56.00542473 +0000 UTC m=+1146.951792545" watchObservedRunningTime="2026-03-20 16:19:56.028293788 +0000 UTC m=+1146.974661603" Mar 20 16:19:56 crc kubenswrapper[4936]: I0320 16:19:56.031214 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-n7kq7" podStartSLOduration=4.143788368 podStartE2EDuration="56.031202483s" podCreationTimestamp="2026-03-20 16:19:00 +0000 UTC" firstStartedPulling="2026-03-20 16:19:01.798933371 +0000 UTC m=+1092.745301186" lastFinishedPulling="2026-03-20 16:19:53.686347486 +0000 UTC m=+1144.632715301" observedRunningTime="2026-03-20 16:19:56.021944132 +0000 UTC m=+1146.968311967" watchObservedRunningTime="2026-03-20 16:19:56.031202483 +0000 UTC m=+1146.977570298" Mar 20 16:19:58 crc kubenswrapper[4936]: I0320 16:19:58.854922 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:19:58 crc kubenswrapper[4936]: I0320 16:19:58.855375 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:19:58 crc kubenswrapper[4936]: I0320 16:19:58.855422 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:19:58 crc kubenswrapper[4936]: I0320 16:19:58.856110 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"929e303b68c86406c3ac9c3e153cfe435d16f382342ffc919886659461c4db6e"} pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 16:19:58 crc kubenswrapper[4936]: I0320 16:19:58.856163 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" containerID="cri-o://929e303b68c86406c3ac9c3e153cfe435d16f382342ffc919886659461c4db6e" gracePeriod=600 Mar 20 16:19:58 crc kubenswrapper[4936]: I0320 16:19:58.999286 4936 generic.go:334] "Generic (PLEG): container finished" podID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerID="929e303b68c86406c3ac9c3e153cfe435d16f382342ffc919886659461c4db6e" exitCode=0 Mar 20 16:19:58 crc kubenswrapper[4936]: I0320 16:19:58.999343 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerDied","Data":"929e303b68c86406c3ac9c3e153cfe435d16f382342ffc919886659461c4db6e"} Mar 20 16:19:58 crc kubenswrapper[4936]: I0320 16:19:58.999378 4936 scope.go:117] "RemoveContainer" containerID="0421fad5e0fd4e5837ad1c312bd061ad68a03b4d2795d6099a681bc4b3ce8f6e" Mar 20 16:19:59 crc kubenswrapper[4936]: I0320 16:19:59.001216 4936 generic.go:334] "Generic (PLEG): container finished" podID="b87c4418-b13e-41e2-bbd8-d7dfb057a594" containerID="198ef6b866a3a9d1b7dc2c7023f673c2fcccad5662c4896ba53b259233660520" exitCode=0 Mar 20 16:19:59 crc kubenswrapper[4936]: I0320 16:19:59.001270 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j4d5s" event={"ID":"b87c4418-b13e-41e2-bbd8-d7dfb057a594","Type":"ContainerDied","Data":"198ef6b866a3a9d1b7dc2c7023f673c2fcccad5662c4896ba53b259233660520"} Mar 20 16:19:59 crc kubenswrapper[4936]: I0320 16:19:59.851262 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-79458c9dc6-vlzgk" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 20 16:19:59 crc kubenswrapper[4936]: I0320 16:19:59.919659 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6ff4f57d48-qp24t" podUID="7f1bd9ee-e0b7-4ea9-b3ca-97987941f149" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.143366 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567060-gbvj5"] Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.159292 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567060-gbvj5" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.162509 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.167076 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567060-gbvj5"] Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.173351 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.174771 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.295654 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pd88\" (UniqueName: \"kubernetes.io/projected/05b79d13-f57b-455c-b5fe-b469c9bc0f45-kube-api-access-2pd88\") pod \"auto-csr-approver-29567060-gbvj5\" (UID: \"05b79d13-f57b-455c-b5fe-b469c9bc0f45\") " pod="openshift-infra/auto-csr-approver-29567060-gbvj5" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.397750 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pd88\" (UniqueName: \"kubernetes.io/projected/05b79d13-f57b-455c-b5fe-b469c9bc0f45-kube-api-access-2pd88\") pod \"auto-csr-approver-29567060-gbvj5\" (UID: \"05b79d13-f57b-455c-b5fe-b469c9bc0f45\") " pod="openshift-infra/auto-csr-approver-29567060-gbvj5" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.418232 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pd88\" (UniqueName: \"kubernetes.io/projected/05b79d13-f57b-455c-b5fe-b469c9bc0f45-kube-api-access-2pd88\") pod \"auto-csr-approver-29567060-gbvj5\" (UID: \"05b79d13-f57b-455c-b5fe-b469c9bc0f45\") " pod="openshift-infra/auto-csr-approver-29567060-gbvj5" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.512782 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567060-gbvj5" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.671611 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j4d5s" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.695999 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.804908 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-config-data\") pod \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.805024 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-scripts\") pod \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.805141 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85cwt\" (UniqueName: \"kubernetes.io/projected/61aff63e-dc1b-4854-af4a-188837ad4cf9-kube-api-access-85cwt\") pod \"61aff63e-dc1b-4854-af4a-188837ad4cf9\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.805192 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g497k\" (UniqueName: \"kubernetes.io/projected/b87c4418-b13e-41e2-bbd8-d7dfb057a594-kube-api-access-g497k\") pod \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.805226 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-db-sync-config-data\") pod \"61aff63e-dc1b-4854-af4a-188837ad4cf9\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.805290 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-combined-ca-bundle\") pod \"61aff63e-dc1b-4854-af4a-188837ad4cf9\" (UID: \"61aff63e-dc1b-4854-af4a-188837ad4cf9\") " Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.805340 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-combined-ca-bundle\") pod \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.806507 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b87c4418-b13e-41e2-bbd8-d7dfb057a594-logs\") pod \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\" (UID: \"b87c4418-b13e-41e2-bbd8-d7dfb057a594\") " Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.807294 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b87c4418-b13e-41e2-bbd8-d7dfb057a594-logs" (OuterVolumeSpecName: "logs") pod "b87c4418-b13e-41e2-bbd8-d7dfb057a594" (UID: "b87c4418-b13e-41e2-bbd8-d7dfb057a594"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.821905 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "61aff63e-dc1b-4854-af4a-188837ad4cf9" (UID: "61aff63e-dc1b-4854-af4a-188837ad4cf9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.824455 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b87c4418-b13e-41e2-bbd8-d7dfb057a594-kube-api-access-g497k" (OuterVolumeSpecName: "kube-api-access-g497k") pod "b87c4418-b13e-41e2-bbd8-d7dfb057a594" (UID: "b87c4418-b13e-41e2-bbd8-d7dfb057a594"). InnerVolumeSpecName "kube-api-access-g497k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.828609 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61aff63e-dc1b-4854-af4a-188837ad4cf9-kube-api-access-85cwt" (OuterVolumeSpecName: "kube-api-access-85cwt") pod "61aff63e-dc1b-4854-af4a-188837ad4cf9" (UID: "61aff63e-dc1b-4854-af4a-188837ad4cf9"). InnerVolumeSpecName "kube-api-access-85cwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.830777 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-scripts" (OuterVolumeSpecName: "scripts") pod "b87c4418-b13e-41e2-bbd8-d7dfb057a594" (UID: "b87c4418-b13e-41e2-bbd8-d7dfb057a594"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.838182 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-config-data" (OuterVolumeSpecName: "config-data") pod "b87c4418-b13e-41e2-bbd8-d7dfb057a594" (UID: "b87c4418-b13e-41e2-bbd8-d7dfb057a594"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.845361 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61aff63e-dc1b-4854-af4a-188837ad4cf9" (UID: "61aff63e-dc1b-4854-af4a-188837ad4cf9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.860869 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b87c4418-b13e-41e2-bbd8-d7dfb057a594" (UID: "b87c4418-b13e-41e2-bbd8-d7dfb057a594"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.908478 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85cwt\" (UniqueName: \"kubernetes.io/projected/61aff63e-dc1b-4854-af4a-188837ad4cf9-kube-api-access-85cwt\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.908568 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g497k\" (UniqueName: \"kubernetes.io/projected/b87c4418-b13e-41e2-bbd8-d7dfb057a594-kube-api-access-g497k\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.908589 4936 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.908602 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61aff63e-dc1b-4854-af4a-188837ad4cf9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.908614 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.908626 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b87c4418-b13e-41e2-bbd8-d7dfb057a594-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.908638 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:00 crc kubenswrapper[4936]: I0320 16:20:00.908649 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b87c4418-b13e-41e2-bbd8-d7dfb057a594-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.023762 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j4d5s" event={"ID":"b87c4418-b13e-41e2-bbd8-d7dfb057a594","Type":"ContainerDied","Data":"241b564a85bfa65256f50f3c4577bf576121d53be0d0ab404e2483ed2f9b616c"} Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.023818 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="241b564a85bfa65256f50f3c4577bf576121d53be0d0ab404e2483ed2f9b616c" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.023897 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j4d5s" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.035472 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9w2gg" event={"ID":"61aff63e-dc1b-4854-af4a-188837ad4cf9","Type":"ContainerDied","Data":"3b77256931847b16a6ddf789a7af732fcc54c4c9b2fff549bc32f63d006f2ba1"} Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.035522 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b77256931847b16a6ddf789a7af732fcc54c4c9b2fff549bc32f63d006f2ba1" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.035530 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9w2gg" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.167693 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.274876 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-584c7d8894-w4d85"] Mar 20 16:20:01 crc kubenswrapper[4936]: E0320 16:20:01.275348 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61aff63e-dc1b-4854-af4a-188837ad4cf9" containerName="barbican-db-sync" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.275372 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="61aff63e-dc1b-4854-af4a-188837ad4cf9" containerName="barbican-db-sync" Mar 20 16:20:01 crc kubenswrapper[4936]: E0320 16:20:01.275407 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b87c4418-b13e-41e2-bbd8-d7dfb057a594" containerName="placement-db-sync" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.275417 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b87c4418-b13e-41e2-bbd8-d7dfb057a594" containerName="placement-db-sync" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.275694 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="61aff63e-dc1b-4854-af4a-188837ad4cf9" containerName="barbican-db-sync" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.275738 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="b87c4418-b13e-41e2-bbd8-d7dfb057a594" containerName="placement-db-sync" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.276863 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.280529 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.280728 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qmxxr" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.280800 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.280883 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.280929 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.293820 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-bnzqf"] Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.294264 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" podUID="ea9e43bd-bedb-48dd-ac3b-bea79379dba0" containerName="dnsmasq-dns" containerID="cri-o://efa25416797404cf3c4d6b6637105ca24d3dfe58e17dfd86e05e9354d2c995a5" gracePeriod=10 Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.305709 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-584c7d8894-w4d85"] Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.420472 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-public-tls-certs\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.420876 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-config-data\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.420917 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-logs\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.420971 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-internal-tls-certs\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.421007 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jpcl\" (UniqueName: \"kubernetes.io/projected/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-kube-api-access-9jpcl\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.421102 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-combined-ca-bundle\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.421145 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-scripts\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.522896 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-combined-ca-bundle\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.523793 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-scripts\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.523845 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-public-tls-certs\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.523909 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-config-data\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.523930 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-logs\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.523969 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-internal-tls-certs\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.524014 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jpcl\" (UniqueName: \"kubernetes.io/projected/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-kube-api-access-9jpcl\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.524945 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-logs\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.527478 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-combined-ca-bundle\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.529086 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-public-tls-certs\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.531104 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-scripts\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.531609 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-internal-tls-certs\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.548863 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-config-data\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.558319 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jpcl\" (UniqueName: \"kubernetes.io/projected/1405ee40-3191-4a1e-9ad7-27b12cd5abaf-kube-api-access-9jpcl\") pod \"placement-584c7d8894-w4d85\" (UID: \"1405ee40-3191-4a1e-9ad7-27b12cd5abaf\") " pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:01 crc kubenswrapper[4936]: I0320 16:20:01.610085 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.020228 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-64855595b6-t5rk7"] Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.021993 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.035340 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.035736 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rpkn6" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.035813 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.049457 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-64855595b6-t5rk7"] Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.068468 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-4tng4"] Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.090513 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.117134 4936 generic.go:334] "Generic (PLEG): container finished" podID="ea9e43bd-bedb-48dd-ac3b-bea79379dba0" containerID="efa25416797404cf3c4d6b6637105ca24d3dfe58e17dfd86e05e9354d2c995a5" exitCode=0 Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.117178 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" event={"ID":"ea9e43bd-bedb-48dd-ac3b-bea79379dba0","Type":"ContainerDied","Data":"efa25416797404cf3c4d6b6637105ca24d3dfe58e17dfd86e05e9354d2c995a5"} Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.130286 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7df458d599-sn4pk"] Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.142339 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.160012 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.183042 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578bb0fa-b068-4f15-9e91-b1ed317b24e6-config-data\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.183126 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plbh7\" (UniqueName: \"kubernetes.io/projected/578bb0fa-b068-4f15-9e91-b1ed317b24e6-kube-api-access-plbh7\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.183176 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.183219 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/578bb0fa-b068-4f15-9e91-b1ed317b24e6-logs\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.183332 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/578bb0fa-b068-4f15-9e91-b1ed317b24e6-config-data-custom\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.183883 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.183923 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-config\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.183950 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578bb0fa-b068-4f15-9e91-b1ed317b24e6-combined-ca-bundle\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.184006 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.184049 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.184065 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf4xw\" (UniqueName: \"kubernetes.io/projected/9a13bb94-e9e5-40bb-b537-29b838225f7b-kube-api-access-gf4xw\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.184156 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-4tng4"] Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.220109 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7df458d599-sn4pk"] Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.267634 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7c9dff68b4-6qmrf"] Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.270080 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.272993 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.275001 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c9dff68b4-6qmrf"] Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.290441 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578bb0fa-b068-4f15-9e91-b1ed317b24e6-config-data\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.290668 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plbh7\" (UniqueName: \"kubernetes.io/projected/578bb0fa-b068-4f15-9e91-b1ed317b24e6-kube-api-access-plbh7\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.290792 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.290890 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/578bb0fa-b068-4f15-9e91-b1ed317b24e6-logs\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.290986 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/578bb0fa-b068-4f15-9e91-b1ed317b24e6-config-data-custom\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.291088 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46c60de1-d091-43da-9f0e-9fa0e31bb556-logs\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.291566 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l25d\" (UniqueName: \"kubernetes.io/projected/46c60de1-d091-43da-9f0e-9fa0e31bb556-kube-api-access-9l25d\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.291662 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46c60de1-d091-43da-9f0e-9fa0e31bb556-config-data-custom\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.291750 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.291822 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-config\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.291896 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578bb0fa-b068-4f15-9e91-b1ed317b24e6-combined-ca-bundle\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.291992 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.292084 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.292155 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf4xw\" (UniqueName: \"kubernetes.io/projected/9a13bb94-e9e5-40bb-b537-29b838225f7b-kube-api-access-gf4xw\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.292241 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c60de1-d091-43da-9f0e-9fa0e31bb556-combined-ca-bundle\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.292327 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c60de1-d091-43da-9f0e-9fa0e31bb556-config-data\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.292087 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/578bb0fa-b068-4f15-9e91-b1ed317b24e6-logs\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.293689 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.294163 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.294226 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.294831 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.294909 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-config\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.304656 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578bb0fa-b068-4f15-9e91-b1ed317b24e6-config-data\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.306496 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/578bb0fa-b068-4f15-9e91-b1ed317b24e6-config-data-custom\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.332754 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578bb0fa-b068-4f15-9e91-b1ed317b24e6-combined-ca-bundle\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.335074 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plbh7\" (UniqueName: \"kubernetes.io/projected/578bb0fa-b068-4f15-9e91-b1ed317b24e6-kube-api-access-plbh7\") pod \"barbican-keystone-listener-64855595b6-t5rk7\" (UID: \"578bb0fa-b068-4f15-9e91-b1ed317b24e6\") " pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.335844 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf4xw\" (UniqueName: \"kubernetes.io/projected/9a13bb94-e9e5-40bb-b537-29b838225f7b-kube-api-access-gf4xw\") pod \"dnsmasq-dns-85ff748b95-4tng4\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.396970 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46c60de1-d091-43da-9f0e-9fa0e31bb556-logs\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.396529 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46c60de1-d091-43da-9f0e-9fa0e31bb556-logs\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.397049 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l25d\" (UniqueName: \"kubernetes.io/projected/46c60de1-d091-43da-9f0e-9fa0e31bb556-kube-api-access-9l25d\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.397069 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46c60de1-d091-43da-9f0e-9fa0e31bb556-config-data-custom\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.397088 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccjj4\" (UniqueName: \"kubernetes.io/projected/802be0f0-c896-47de-8efa-24c499ed3e2c-kube-api-access-ccjj4\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.397112 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-combined-ca-bundle\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.397187 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.397223 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802be0f0-c896-47de-8efa-24c499ed3e2c-logs\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.397280 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c60de1-d091-43da-9f0e-9fa0e31bb556-combined-ca-bundle\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.397309 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data-custom\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.397329 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c60de1-d091-43da-9f0e-9fa0e31bb556-config-data\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.401747 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c60de1-d091-43da-9f0e-9fa0e31bb556-config-data\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.407524 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c60de1-d091-43da-9f0e-9fa0e31bb556-combined-ca-bundle\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.412405 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46c60de1-d091-43da-9f0e-9fa0e31bb556-config-data-custom\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.417062 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.424182 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l25d\" (UniqueName: \"kubernetes.io/projected/46c60de1-d091-43da-9f0e-9fa0e31bb556-kube-api-access-9l25d\") pod \"barbican-worker-7df458d599-sn4pk\" (UID: \"46c60de1-d091-43da-9f0e-9fa0e31bb556\") " pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.484490 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.504581 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccjj4\" (UniqueName: \"kubernetes.io/projected/802be0f0-c896-47de-8efa-24c499ed3e2c-kube-api-access-ccjj4\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.504689 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-combined-ca-bundle\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.504958 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.507557 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802be0f0-c896-47de-8efa-24c499ed3e2c-logs\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.507887 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data-custom\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.509601 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802be0f0-c896-47de-8efa-24c499ed3e2c-logs\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.513099 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.513711 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-combined-ca-bundle\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.519004 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data-custom\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.526147 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccjj4\" (UniqueName: \"kubernetes.io/projected/802be0f0-c896-47de-8efa-24c499ed3e2c-kube-api-access-ccjj4\") pod \"barbican-api-7c9dff68b4-6qmrf\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.526686 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7df458d599-sn4pk" Mar 20 16:20:02 crc kubenswrapper[4936]: I0320 16:20:02.597116 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.176499 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.189448 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" event={"ID":"ea9e43bd-bedb-48dd-ac3b-bea79379dba0","Type":"ContainerDied","Data":"3900f92575c8232b772ffd98b70ce2de539b2b95db78a10e89004e5b92b4ee3d"} Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.189510 4936 scope.go:117] "RemoveContainer" containerID="efa25416797404cf3c4d6b6637105ca24d3dfe58e17dfd86e05e9354d2c995a5" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.282597 4936 scope.go:117] "RemoveContainer" containerID="636c1037b4c4e3eae5558e4631e0729053e7fb4b012927f4e120352c40cba7bf" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.355273 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-nb\") pod \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.355831 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhgx4\" (UniqueName: \"kubernetes.io/projected/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-kube-api-access-bhgx4\") pod \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.355895 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-sb\") pod \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.356025 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-svc\") pod \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.356052 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-swift-storage-0\") pod \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.356187 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-config\") pod \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\" (UID: \"ea9e43bd-bedb-48dd-ac3b-bea79379dba0\") " Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.371818 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-kube-api-access-bhgx4" (OuterVolumeSpecName: "kube-api-access-bhgx4") pod "ea9e43bd-bedb-48dd-ac3b-bea79379dba0" (UID: "ea9e43bd-bedb-48dd-ac3b-bea79379dba0"). InnerVolumeSpecName "kube-api-access-bhgx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.468481 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhgx4\" (UniqueName: \"kubernetes.io/projected/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-kube-api-access-bhgx4\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:04 crc kubenswrapper[4936]: E0320 16:20:04.522673 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="052101db-ddd5-419e-8131-fccbe44241b4" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.538532 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea9e43bd-bedb-48dd-ac3b-bea79379dba0" (UID: "ea9e43bd-bedb-48dd-ac3b-bea79379dba0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.544491 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-config" (OuterVolumeSpecName: "config") pod "ea9e43bd-bedb-48dd-ac3b-bea79379dba0" (UID: "ea9e43bd-bedb-48dd-ac3b-bea79379dba0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.549418 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea9e43bd-bedb-48dd-ac3b-bea79379dba0" (UID: "ea9e43bd-bedb-48dd-ac3b-bea79379dba0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.551834 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ea9e43bd-bedb-48dd-ac3b-bea79379dba0" (UID: "ea9e43bd-bedb-48dd-ac3b-bea79379dba0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.551872 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea9e43bd-bedb-48dd-ac3b-bea79379dba0" (UID: "ea9e43bd-bedb-48dd-ac3b-bea79379dba0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.570339 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.570379 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.570391 4936 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.570401 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.570413 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea9e43bd-bedb-48dd-ac3b-bea79379dba0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.940648 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567060-gbvj5"] Mar 20 16:20:04 crc kubenswrapper[4936]: W0320 16:20:04.958235 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod802be0f0_c896_47de_8efa_24c499ed3e2c.slice/crio-45ee1d9e85d6473da389796355687f44adc633538e08ce4f344dd54321d7c785 WatchSource:0}: Error finding container 45ee1d9e85d6473da389796355687f44adc633538e08ce4f344dd54321d7c785: Status 404 returned error can't find the container with id 45ee1d9e85d6473da389796355687f44adc633538e08ce4f344dd54321d7c785 Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.963963 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c9dff68b4-6qmrf"] Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.972031 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-4tng4"] Mar 20 16:20:04 crc kubenswrapper[4936]: W0320 16:20:04.975993 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1405ee40_3191_4a1e_9ad7_27b12cd5abaf.slice/crio-a61bc4be318c2aac488203b1df3f0902d862ac8cba5d88f12b61f92a5485114a WatchSource:0}: Error finding container a61bc4be318c2aac488203b1df3f0902d862ac8cba5d88f12b61f92a5485114a: Status 404 returned error can't find the container with id a61bc4be318c2aac488203b1df3f0902d862ac8cba5d88f12b61f92a5485114a Mar 20 16:20:04 crc kubenswrapper[4936]: I0320 16:20:04.984325 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-584c7d8894-w4d85"] Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.021854 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7df458d599-sn4pk"] Mar 20 16:20:05 crc kubenswrapper[4936]: W0320 16:20:05.045060 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46c60de1_d091_43da_9f0e_9fa0e31bb556.slice/crio-4324ebc097d2a12477a2d8df6edc914cd54c00996350fbd9b5edfae8d7c279c4 WatchSource:0}: Error finding container 4324ebc097d2a12477a2d8df6edc914cd54c00996350fbd9b5edfae8d7c279c4: Status 404 returned error can't find the container with id 4324ebc097d2a12477a2d8df6edc914cd54c00996350fbd9b5edfae8d7c279c4 Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.103413 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-64855595b6-t5rk7"] Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.170914 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-598cbc676b-tm5b2"] Mar 20 16:20:05 crc kubenswrapper[4936]: E0320 16:20:05.174378 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9e43bd-bedb-48dd-ac3b-bea79379dba0" containerName="init" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.174439 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9e43bd-bedb-48dd-ac3b-bea79379dba0" containerName="init" Mar 20 16:20:05 crc kubenswrapper[4936]: E0320 16:20:05.174505 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9e43bd-bedb-48dd-ac3b-bea79379dba0" containerName="dnsmasq-dns" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.174518 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9e43bd-bedb-48dd-ac3b-bea79379dba0" containerName="dnsmasq-dns" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.187471 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea9e43bd-bedb-48dd-ac3b-bea79379dba0" containerName="dnsmasq-dns" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.188423 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.194008 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.194196 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.198501 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-598cbc676b-tm5b2"] Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.217666 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"052101db-ddd5-419e-8131-fccbe44241b4","Type":"ContainerStarted","Data":"bddd4747dc679ad51d79a0cb995b7c9a2e57bc05d68940e8a53328a45b926418"} Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.217841 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="ceilometer-notification-agent" containerID="cri-o://38e37211aa92495ebb6cf8608eefe13a16c2bebaa4f1fcedc52ab95d9fa21412" gracePeriod=30 Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.218077 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.220651 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="proxy-httpd" containerID="cri-o://bddd4747dc679ad51d79a0cb995b7c9a2e57bc05d68940e8a53328a45b926418" gracePeriod=30 Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.220684 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="sg-core" containerID="cri-o://d672b85638441f903d415abe7efb0ea9860db5fac56d38f39f99b39e37d243f7" gracePeriod=30 Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.239562 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567060-gbvj5" event={"ID":"05b79d13-f57b-455c-b5fe-b469c9bc0f45","Type":"ContainerStarted","Data":"ffcef1ebbee1dc9cd7e0967deb1fccb2df0516b28720cce9060156b29956c1cd"} Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.250746 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" event={"ID":"9a13bb94-e9e5-40bb-b537-29b838225f7b","Type":"ContainerStarted","Data":"a4007f9cdabf76a6e57b4009ac9d167043b9a77342f2ec00e580cc3232b39cf1"} Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.263718 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-bnzqf" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.270516 4936 generic.go:334] "Generic (PLEG): container finished" podID="047fb63e-64a8-4280-9846-659575038df3" containerID="35818d37873ba37611214468b71d71e9107f339eac76ba0c3a162bac351a8d84" exitCode=0 Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.270600 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n7kq7" event={"ID":"047fb63e-64a8-4280-9846-659575038df3","Type":"ContainerDied","Data":"35818d37873ba37611214468b71d71e9107f339eac76ba0c3a162bac351a8d84"} Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.274177 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7df458d599-sn4pk" event={"ID":"46c60de1-d091-43da-9f0e-9fa0e31bb556","Type":"ContainerStarted","Data":"4324ebc097d2a12477a2d8df6edc914cd54c00996350fbd9b5edfae8d7c279c4"} Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.275292 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" event={"ID":"578bb0fa-b068-4f15-9e91-b1ed317b24e6","Type":"ContainerStarted","Data":"87abe78da16a0d9e8f78a1f89d0b528bd655af575b1285caa0b2a3b58ca3b5d4"} Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.280664 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"19b7ae40534ebc0df1ca16d7a38e73567bfe3ddd0d163640d422629548acfbe8"} Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.283010 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c9dff68b4-6qmrf" event={"ID":"802be0f0-c896-47de-8efa-24c499ed3e2c","Type":"ContainerStarted","Data":"45ee1d9e85d6473da389796355687f44adc633538e08ce4f344dd54321d7c785"} Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.285034 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-584c7d8894-w4d85" event={"ID":"1405ee40-3191-4a1e-9ad7-27b12cd5abaf","Type":"ContainerStarted","Data":"a61bc4be318c2aac488203b1df3f0902d862ac8cba5d88f12b61f92a5485114a"} Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.307181 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-config-data-custom\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.307243 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-combined-ca-bundle\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.307283 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-public-tls-certs\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.307361 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-config-data\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.307420 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-logs\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.307482 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-internal-tls-certs\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.307578 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7njk\" (UniqueName: \"kubernetes.io/projected/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-kube-api-access-p7njk\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.310570 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-bnzqf"] Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.323243 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-bnzqf"] Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.409238 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-internal-tls-certs\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.410176 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7njk\" (UniqueName: \"kubernetes.io/projected/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-kube-api-access-p7njk\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.410362 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-config-data-custom\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.410530 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-combined-ca-bundle\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.410684 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-public-tls-certs\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.410828 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-config-data\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.410959 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-logs\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.412901 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-logs\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.414464 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-internal-tls-certs\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.414705 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-config-data-custom\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.417926 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-combined-ca-bundle\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.418089 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-public-tls-certs\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.423030 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-config-data\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.430847 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7njk\" (UniqueName: \"kubernetes.io/projected/de8a5591-dee9-46b2-81d8-8cfa2c8dce57-kube-api-access-p7njk\") pod \"barbican-api-598cbc676b-tm5b2\" (UID: \"de8a5591-dee9-46b2-81d8-8cfa2c8dce57\") " pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.516431 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:05 crc kubenswrapper[4936]: I0320 16:20:05.872851 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea9e43bd-bedb-48dd-ac3b-bea79379dba0" path="/var/lib/kubelet/pods/ea9e43bd-bedb-48dd-ac3b-bea79379dba0/volumes" Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.015224 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-598cbc676b-tm5b2"] Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.318510 4936 generic.go:334] "Generic (PLEG): container finished" podID="052101db-ddd5-419e-8131-fccbe44241b4" containerID="bddd4747dc679ad51d79a0cb995b7c9a2e57bc05d68940e8a53328a45b926418" exitCode=0 Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.318828 4936 generic.go:334] "Generic (PLEG): container finished" podID="052101db-ddd5-419e-8131-fccbe44241b4" containerID="d672b85638441f903d415abe7efb0ea9860db5fac56d38f39f99b39e37d243f7" exitCode=2 Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.318591 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"052101db-ddd5-419e-8131-fccbe44241b4","Type":"ContainerDied","Data":"bddd4747dc679ad51d79a0cb995b7c9a2e57bc05d68940e8a53328a45b926418"} Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.318954 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"052101db-ddd5-419e-8131-fccbe44241b4","Type":"ContainerDied","Data":"d672b85638441f903d415abe7efb0ea9860db5fac56d38f39f99b39e37d243f7"} Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.326149 4936 generic.go:334] "Generic (PLEG): container finished" podID="9a13bb94-e9e5-40bb-b537-29b838225f7b" containerID="14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec" exitCode=0 Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.326931 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" event={"ID":"9a13bb94-e9e5-40bb-b537-29b838225f7b","Type":"ContainerDied","Data":"14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec"} Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.330626 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-598cbc676b-tm5b2" event={"ID":"de8a5591-dee9-46b2-81d8-8cfa2c8dce57","Type":"ContainerStarted","Data":"110a4c4f476d860aa429a7d5d098fb7553a9b56477063302d1f5b34c11c04726"} Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.330711 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-598cbc676b-tm5b2" event={"ID":"de8a5591-dee9-46b2-81d8-8cfa2c8dce57","Type":"ContainerStarted","Data":"8f05728411aa9f01d5a16c143785163f516511536044078e2f5e113c0747aff0"} Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.336331 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-584c7d8894-w4d85" event={"ID":"1405ee40-3191-4a1e-9ad7-27b12cd5abaf","Type":"ContainerStarted","Data":"632fe5ac3d4110f6598002d018c0ca93b85d6438a63a3d64372180ee40313150"} Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.336387 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-584c7d8894-w4d85" event={"ID":"1405ee40-3191-4a1e-9ad7-27b12cd5abaf","Type":"ContainerStarted","Data":"84d5f0f4647e83d198170279a69544effdaa262bd485bd7dbde78f29bd55c4bd"} Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.336412 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.336491 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.340667 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c9dff68b4-6qmrf" event={"ID":"802be0f0-c896-47de-8efa-24c499ed3e2c","Type":"ContainerStarted","Data":"656380c7600e02c0c0e14072bc9a0e44503c8ea8975917d7d4d207dec2f80f4f"} Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.340705 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c9dff68b4-6qmrf" event={"ID":"802be0f0-c896-47de-8efa-24c499ed3e2c","Type":"ContainerStarted","Data":"cddaadd66f02ccb8e7888193cf6b62fe6fc34d4566ea8e87c4a4fbd5b3c5d6a9"} Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.340719 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.340839 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.390162 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-584c7d8894-w4d85" podStartSLOduration=5.390128035 podStartE2EDuration="5.390128035s" podCreationTimestamp="2026-03-20 16:20:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:06.378571258 +0000 UTC m=+1157.324939073" watchObservedRunningTime="2026-03-20 16:20:06.390128035 +0000 UTC m=+1157.336495850" Mar 20 16:20:06 crc kubenswrapper[4936]: I0320 16:20:06.407410 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7c9dff68b4-6qmrf" podStartSLOduration=4.407388829 podStartE2EDuration="4.407388829s" podCreationTimestamp="2026-03-20 16:20:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:06.403884167 +0000 UTC m=+1157.350251982" watchObservedRunningTime="2026-03-20 16:20:06.407388829 +0000 UTC m=+1157.353756634" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.112679 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.142657 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-combined-ca-bundle\") pod \"047fb63e-64a8-4280-9846-659575038df3\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.142787 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-scripts\") pod \"047fb63e-64a8-4280-9846-659575038df3\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.142851 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-db-sync-config-data\") pod \"047fb63e-64a8-4280-9846-659575038df3\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.142884 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/047fb63e-64a8-4280-9846-659575038df3-etc-machine-id\") pod \"047fb63e-64a8-4280-9846-659575038df3\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.142915 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-config-data\") pod \"047fb63e-64a8-4280-9846-659575038df3\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.143055 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j49gd\" (UniqueName: \"kubernetes.io/projected/047fb63e-64a8-4280-9846-659575038df3-kube-api-access-j49gd\") pod \"047fb63e-64a8-4280-9846-659575038df3\" (UID: \"047fb63e-64a8-4280-9846-659575038df3\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.143786 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/047fb63e-64a8-4280-9846-659575038df3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "047fb63e-64a8-4280-9846-659575038df3" (UID: "047fb63e-64a8-4280-9846-659575038df3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.149429 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-scripts" (OuterVolumeSpecName: "scripts") pod "047fb63e-64a8-4280-9846-659575038df3" (UID: "047fb63e-64a8-4280-9846-659575038df3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.149441 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/047fb63e-64a8-4280-9846-659575038df3-kube-api-access-j49gd" (OuterVolumeSpecName: "kube-api-access-j49gd") pod "047fb63e-64a8-4280-9846-659575038df3" (UID: "047fb63e-64a8-4280-9846-659575038df3"). InnerVolumeSpecName "kube-api-access-j49gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.149470 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "047fb63e-64a8-4280-9846-659575038df3" (UID: "047fb63e-64a8-4280-9846-659575038df3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.178469 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "047fb63e-64a8-4280-9846-659575038df3" (UID: "047fb63e-64a8-4280-9846-659575038df3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.224322 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-config-data" (OuterVolumeSpecName: "config-data") pod "047fb63e-64a8-4280-9846-659575038df3" (UID: "047fb63e-64a8-4280-9846-659575038df3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.245487 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.245531 4936 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.245568 4936 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/047fb63e-64a8-4280-9846-659575038df3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.245579 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.245591 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j49gd\" (UniqueName: \"kubernetes.io/projected/047fb63e-64a8-4280-9846-659575038df3-kube-api-access-j49gd\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.245602 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047fb63e-64a8-4280-9846-659575038df3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.349755 4936 generic.go:334] "Generic (PLEG): container finished" podID="052101db-ddd5-419e-8131-fccbe44241b4" containerID="38e37211aa92495ebb6cf8608eefe13a16c2bebaa4f1fcedc52ab95d9fa21412" exitCode=0 Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.349808 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"052101db-ddd5-419e-8131-fccbe44241b4","Type":"ContainerDied","Data":"38e37211aa92495ebb6cf8608eefe13a16c2bebaa4f1fcedc52ab95d9fa21412"} Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.351780 4936 generic.go:334] "Generic (PLEG): container finished" podID="05b79d13-f57b-455c-b5fe-b469c9bc0f45" containerID="bc0e47a1192a0d56c16fbbc84a264cfcc6b0bcb470db43b06e0788b81d26b633" exitCode=0 Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.352262 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567060-gbvj5" event={"ID":"05b79d13-f57b-455c-b5fe-b469c9bc0f45","Type":"ContainerDied","Data":"bc0e47a1192a0d56c16fbbc84a264cfcc6b0bcb470db43b06e0788b81d26b633"} Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.355452 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" event={"ID":"9a13bb94-e9e5-40bb-b537-29b838225f7b","Type":"ContainerStarted","Data":"8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726"} Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.355912 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.360131 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-598cbc676b-tm5b2" event={"ID":"de8a5591-dee9-46b2-81d8-8cfa2c8dce57","Type":"ContainerStarted","Data":"07d3fe99be345234f3931dedca9ea391ed954ae30c851d5d5d4ea6b28f6a2846"} Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.360227 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.360272 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.370163 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n7kq7" event={"ID":"047fb63e-64a8-4280-9846-659575038df3","Type":"ContainerDied","Data":"e8d55bc7390adcfd9d73f556db8df5783b0e984ca6753ed85c82967c7bace81d"} Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.370204 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n7kq7" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.370209 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8d55bc7390adcfd9d73f556db8df5783b0e984ca6753ed85c82967c7bace81d" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.386815 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" podStartSLOduration=5.38678595 podStartE2EDuration="5.38678595s" podCreationTimestamp="2026-03-20 16:20:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:07.384289077 +0000 UTC m=+1158.330656912" watchObservedRunningTime="2026-03-20 16:20:07.38678595 +0000 UTC m=+1158.333153765" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.409753 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-598cbc676b-tm5b2" podStartSLOduration=2.4097317289999998 podStartE2EDuration="2.409731729s" podCreationTimestamp="2026-03-20 16:20:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:07.408533764 +0000 UTC m=+1158.354901599" watchObservedRunningTime="2026-03-20 16:20:07.409731729 +0000 UTC m=+1158.356099544" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.576165 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.636071 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 20 16:20:07 crc kubenswrapper[4936]: E0320 16:20:07.636495 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="ceilometer-notification-agent" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.636511 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="ceilometer-notification-agent" Mar 20 16:20:07 crc kubenswrapper[4936]: E0320 16:20:07.636563 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="proxy-httpd" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.636573 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="proxy-httpd" Mar 20 16:20:07 crc kubenswrapper[4936]: E0320 16:20:07.636599 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047fb63e-64a8-4280-9846-659575038df3" containerName="cinder-db-sync" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.636608 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="047fb63e-64a8-4280-9846-659575038df3" containerName="cinder-db-sync" Mar 20 16:20:07 crc kubenswrapper[4936]: E0320 16:20:07.636628 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="sg-core" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.636636 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="sg-core" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.636858 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="047fb63e-64a8-4280-9846-659575038df3" containerName="cinder-db-sync" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.636884 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="sg-core" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.636899 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="proxy-httpd" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.636924 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="052101db-ddd5-419e-8131-fccbe44241b4" containerName="ceilometer-notification-agent" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.649291 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.657662 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-scripts\") pod \"052101db-ddd5-419e-8131-fccbe44241b4\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.658811 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-config-data\") pod \"052101db-ddd5-419e-8131-fccbe44241b4\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.658955 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrdhw\" (UniqueName: \"kubernetes.io/projected/052101db-ddd5-419e-8131-fccbe44241b4-kube-api-access-wrdhw\") pod \"052101db-ddd5-419e-8131-fccbe44241b4\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.659013 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-sg-core-conf-yaml\") pod \"052101db-ddd5-419e-8131-fccbe44241b4\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.659063 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-combined-ca-bundle\") pod \"052101db-ddd5-419e-8131-fccbe44241b4\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.659113 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-log-httpd\") pod \"052101db-ddd5-419e-8131-fccbe44241b4\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.659135 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-run-httpd\") pod \"052101db-ddd5-419e-8131-fccbe44241b4\" (UID: \"052101db-ddd5-419e-8131-fccbe44241b4\") " Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.665917 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "052101db-ddd5-419e-8131-fccbe44241b4" (UID: "052101db-ddd5-419e-8131-fccbe44241b4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.666799 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "052101db-ddd5-419e-8131-fccbe44241b4" (UID: "052101db-ddd5-419e-8131-fccbe44241b4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.671950 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-scripts" (OuterVolumeSpecName: "scripts") pod "052101db-ddd5-419e-8131-fccbe44241b4" (UID: "052101db-ddd5-419e-8131-fccbe44241b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.672461 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.672498 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.672747 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7kz5m" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.673174 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.708065 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/052101db-ddd5-419e-8131-fccbe44241b4-kube-api-access-wrdhw" (OuterVolumeSpecName: "kube-api-access-wrdhw") pod "052101db-ddd5-419e-8131-fccbe44241b4" (UID: "052101db-ddd5-419e-8131-fccbe44241b4"). InnerVolumeSpecName "kube-api-access-wrdhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.721754 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.760778 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.760861 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-scripts\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.760944 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.760981 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.761020 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzdst\" (UniqueName: \"kubernetes.io/projected/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-kube-api-access-nzdst\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.761051 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.761132 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.761147 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrdhw\" (UniqueName: \"kubernetes.io/projected/052101db-ddd5-419e-8131-fccbe44241b4-kube-api-access-wrdhw\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.761158 4936 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.761167 4936 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/052101db-ddd5-419e-8131-fccbe44241b4-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.771704 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-4tng4"] Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.780787 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "052101db-ddd5-419e-8131-fccbe44241b4" (UID: "052101db-ddd5-419e-8131-fccbe44241b4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.790731 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lndxt"] Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.793084 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.797310 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lndxt"] Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.819718 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "052101db-ddd5-419e-8131-fccbe44241b4" (UID: "052101db-ddd5-419e-8131-fccbe44241b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862225 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862282 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-scripts\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862327 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-config\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862356 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862388 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862418 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sr8t\" (UniqueName: \"kubernetes.io/projected/b2a56dab-8614-4f51-bbe7-46f59dda60de-kube-api-access-4sr8t\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862434 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862475 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzdst\" (UniqueName: \"kubernetes.io/projected/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-kube-api-access-nzdst\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862494 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862516 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862535 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862639 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862730 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.862742 4936 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.863747 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.867203 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.873966 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.876435 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.879642 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-config-data" (OuterVolumeSpecName: "config-data") pod "052101db-ddd5-419e-8131-fccbe44241b4" (UID: "052101db-ddd5-419e-8131-fccbe44241b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.881844 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-scripts\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.889720 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzdst\" (UniqueName: \"kubernetes.io/projected/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-kube-api-access-nzdst\") pod \"cinder-scheduler-0\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.943511 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.945145 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.945238 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.946984 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.963937 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-config\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.963998 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.964077 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sr8t\" (UniqueName: \"kubernetes.io/projected/b2a56dab-8614-4f51-bbe7-46f59dda60de-kube-api-access-4sr8t\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.965168 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.965216 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-config\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.965316 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.966242 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.967562 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.967639 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.967839 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052101db-ddd5-419e-8131-fccbe44241b4-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.967988 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.968034 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:07 crc kubenswrapper[4936]: I0320 16:20:07.971102 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.001626 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sr8t\" (UniqueName: \"kubernetes.io/projected/b2a56dab-8614-4f51-bbe7-46f59dda60de-kube-api-access-4sr8t\") pod \"dnsmasq-dns-5c9776ccc5-lndxt\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.069947 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.070018 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djkk2\" (UniqueName: \"kubernetes.io/projected/ff274176-aa07-4573-b1f3-a6fa882b08b5-kube-api-access-djkk2\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.070050 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.070074 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff274176-aa07-4573-b1f3-a6fa882b08b5-logs\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.070158 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.070222 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff274176-aa07-4573-b1f3-a6fa882b08b5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.070266 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-scripts\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.171483 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.171580 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff274176-aa07-4573-b1f3-a6fa882b08b5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.171629 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-scripts\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.171731 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.171727 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff274176-aa07-4573-b1f3-a6fa882b08b5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.171766 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djkk2\" (UniqueName: \"kubernetes.io/projected/ff274176-aa07-4573-b1f3-a6fa882b08b5-kube-api-access-djkk2\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.171832 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.171868 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff274176-aa07-4573-b1f3-a6fa882b08b5-logs\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.172493 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff274176-aa07-4573-b1f3-a6fa882b08b5-logs\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.176349 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-scripts\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.176635 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.177398 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.178269 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.191741 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djkk2\" (UniqueName: \"kubernetes.io/projected/ff274176-aa07-4573-b1f3-a6fa882b08b5-kube-api-access-djkk2\") pod \"cinder-api-0\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.290981 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.363252 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.386859 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7df458d599-sn4pk" event={"ID":"46c60de1-d091-43da-9f0e-9fa0e31bb556","Type":"ContainerStarted","Data":"ba58d85c9989c4231c12458b2d647604c51f1b3149424850eb6d5722c07b648f"} Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.386903 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7df458d599-sn4pk" event={"ID":"46c60de1-d091-43da-9f0e-9fa0e31bb556","Type":"ContainerStarted","Data":"3beeda3e1641debab2770d4fac8963894344cde7cb8c8fdabf531c35bb13a0a1"} Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.397764 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" event={"ID":"578bb0fa-b068-4f15-9e91-b1ed317b24e6","Type":"ContainerStarted","Data":"5341a80c6842221ca7761cd7b5ce120385d8a5bb227c71109f36436af8df9242"} Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.397809 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" event={"ID":"578bb0fa-b068-4f15-9e91-b1ed317b24e6","Type":"ContainerStarted","Data":"d067add72b459d27ac00f6512185abbde0ad7324f30de53e6e652acc622c7817"} Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.411042 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"052101db-ddd5-419e-8131-fccbe44241b4","Type":"ContainerDied","Data":"8a346d219c8a9854c8e7822f48f552381141090bce1735a9d41237dffbb7d512"} Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.411085 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.411100 4936 scope.go:117] "RemoveContainer" containerID="bddd4747dc679ad51d79a0cb995b7c9a2e57bc05d68940e8a53328a45b926418" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.424234 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7df458d599-sn4pk" podStartSLOduration=4.893150634 podStartE2EDuration="7.424211804s" podCreationTimestamp="2026-03-20 16:20:01 +0000 UTC" firstStartedPulling="2026-03-20 16:20:05.065499551 +0000 UTC m=+1156.011867366" lastFinishedPulling="2026-03-20 16:20:07.596560721 +0000 UTC m=+1158.542928536" observedRunningTime="2026-03-20 16:20:08.405519549 +0000 UTC m=+1159.351887364" watchObservedRunningTime="2026-03-20 16:20:08.424211804 +0000 UTC m=+1159.370579619" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.499591 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-64855595b6-t5rk7" podStartSLOduration=4.9667223499999995 podStartE2EDuration="7.499565853s" podCreationTimestamp="2026-03-20 16:20:01 +0000 UTC" firstStartedPulling="2026-03-20 16:20:05.057224369 +0000 UTC m=+1156.003592184" lastFinishedPulling="2026-03-20 16:20:07.590067872 +0000 UTC m=+1158.536435687" observedRunningTime="2026-03-20 16:20:08.427117569 +0000 UTC m=+1159.373485384" watchObservedRunningTime="2026-03-20 16:20:08.499565853 +0000 UTC m=+1159.445933668" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.525689 4936 scope.go:117] "RemoveContainer" containerID="d672b85638441f903d415abe7efb0ea9860db5fac56d38f39f99b39e37d243f7" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.538940 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.569286 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.579744 4936 scope.go:117] "RemoveContainer" containerID="38e37211aa92495ebb6cf8608eefe13a16c2bebaa4f1fcedc52ab95d9fa21412" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.589032 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.605227 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.609180 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.611283 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.611635 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.639836 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.683154 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-log-httpd\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.683244 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.683274 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-config-data\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.683303 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-scripts\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.683326 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-run-httpd\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.683352 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.683401 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b4br\" (UniqueName: \"kubernetes.io/projected/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-kube-api-access-2b4br\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.784622 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b4br\" (UniqueName: \"kubernetes.io/projected/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-kube-api-access-2b4br\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.785196 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-log-httpd\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.785292 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.785319 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-config-data\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.785349 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-scripts\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.785371 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-run-httpd\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.785400 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.788412 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-log-httpd\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.789568 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-run-httpd\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.795334 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-scripts\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.797338 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.805687 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lndxt"] Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.806520 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.810338 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b4br\" (UniqueName: \"kubernetes.io/projected/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-kube-api-access-2b4br\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.810812 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-config-data\") pod \"ceilometer-0\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " pod="openstack/ceilometer-0" Mar 20 16:20:08 crc kubenswrapper[4936]: I0320 16:20:08.936111 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.014307 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567060-gbvj5" Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.090198 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pd88\" (UniqueName: \"kubernetes.io/projected/05b79d13-f57b-455c-b5fe-b469c9bc0f45-kube-api-access-2pd88\") pod \"05b79d13-f57b-455c-b5fe-b469c9bc0f45\" (UID: \"05b79d13-f57b-455c-b5fe-b469c9bc0f45\") " Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.102408 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05b79d13-f57b-455c-b5fe-b469c9bc0f45-kube-api-access-2pd88" (OuterVolumeSpecName: "kube-api-access-2pd88") pod "05b79d13-f57b-455c-b5fe-b469c9bc0f45" (UID: "05b79d13-f57b-455c-b5fe-b469c9bc0f45"). InnerVolumeSpecName "kube-api-access-2pd88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:09 crc kubenswrapper[4936]: W0320 16:20:09.158808 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff274176_aa07_4573_b1f3_a6fa882b08b5.slice/crio-7d8fcb4839a638e8fe82b92955a7ec07897585cb29517cedf7fe76effb39e9e0 WatchSource:0}: Error finding container 7d8fcb4839a638e8fe82b92955a7ec07897585cb29517cedf7fe76effb39e9e0: Status 404 returned error can't find the container with id 7d8fcb4839a638e8fe82b92955a7ec07897585cb29517cedf7fe76effb39e9e0 Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.173291 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.200514 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pd88\" (UniqueName: \"kubernetes.io/projected/05b79d13-f57b-455c-b5fe-b469c9bc0f45-kube-api-access-2pd88\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.459448 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c","Type":"ContainerStarted","Data":"46a76fee5b45fa1fb21d40c15d111803c7a5b5057f2ac9c1fb794da93e1caddc"} Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.477219 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff274176-aa07-4573-b1f3-a6fa882b08b5","Type":"ContainerStarted","Data":"7d8fcb4839a638e8fe82b92955a7ec07897585cb29517cedf7fe76effb39e9e0"} Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.507106 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.507585 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567060-gbvj5" event={"ID":"05b79d13-f57b-455c-b5fe-b469c9bc0f45","Type":"ContainerDied","Data":"ffcef1ebbee1dc9cd7e0967deb1fccb2df0516b28720cce9060156b29956c1cd"} Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.507614 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffcef1ebbee1dc9cd7e0967deb1fccb2df0516b28720cce9060156b29956c1cd" Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.507682 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567060-gbvj5" Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.526423 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" event={"ID":"b2a56dab-8614-4f51-bbe7-46f59dda60de","Type":"ContainerStarted","Data":"ca32faf9fcc7baa65af0e7c5ef5eb0a4a30b2e8d5d202bc7701f2b08d53b9ee9"} Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.526472 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" event={"ID":"b2a56dab-8614-4f51-bbe7-46f59dda60de","Type":"ContainerStarted","Data":"ec0a4f6d86ce6d6a243d3bd4a3022259c8c960ffe21ecc6d7702082c7f2669f4"} Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.527428 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" podUID="9a13bb94-e9e5-40bb-b537-29b838225f7b" containerName="dnsmasq-dns" containerID="cri-o://8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726" gracePeriod=10 Mar 20 16:20:09 crc kubenswrapper[4936]: I0320 16:20:09.888870 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="052101db-ddd5-419e-8131-fccbe44241b4" path="/var/lib/kubelet/pods/052101db-ddd5-419e-8131-fccbe44241b4/volumes" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.035171 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.115748 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567054-xkcjv"] Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.150032 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567054-xkcjv"] Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.293198 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.388567 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-swift-storage-0\") pod \"9a13bb94-e9e5-40bb-b537-29b838225f7b\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.388624 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-nb\") pod \"9a13bb94-e9e5-40bb-b537-29b838225f7b\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.388716 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-svc\") pod \"9a13bb94-e9e5-40bb-b537-29b838225f7b\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.388785 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-sb\") pod \"9a13bb94-e9e5-40bb-b537-29b838225f7b\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.388860 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-config\") pod \"9a13bb94-e9e5-40bb-b537-29b838225f7b\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.388982 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf4xw\" (UniqueName: \"kubernetes.io/projected/9a13bb94-e9e5-40bb-b537-29b838225f7b-kube-api-access-gf4xw\") pod \"9a13bb94-e9e5-40bb-b537-29b838225f7b\" (UID: \"9a13bb94-e9e5-40bb-b537-29b838225f7b\") " Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.403560 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a13bb94-e9e5-40bb-b537-29b838225f7b-kube-api-access-gf4xw" (OuterVolumeSpecName: "kube-api-access-gf4xw") pod "9a13bb94-e9e5-40bb-b537-29b838225f7b" (UID: "9a13bb94-e9e5-40bb-b537-29b838225f7b"). InnerVolumeSpecName "kube-api-access-gf4xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.440199 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a13bb94-e9e5-40bb-b537-29b838225f7b" (UID: "9a13bb94-e9e5-40bb-b537-29b838225f7b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.448894 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-config" (OuterVolumeSpecName: "config") pod "9a13bb94-e9e5-40bb-b537-29b838225f7b" (UID: "9a13bb94-e9e5-40bb-b537-29b838225f7b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.449908 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9a13bb94-e9e5-40bb-b537-29b838225f7b" (UID: "9a13bb94-e9e5-40bb-b537-29b838225f7b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.464805 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9a13bb94-e9e5-40bb-b537-29b838225f7b" (UID: "9a13bb94-e9e5-40bb-b537-29b838225f7b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.481496 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9a13bb94-e9e5-40bb-b537-29b838225f7b" (UID: "9a13bb94-e9e5-40bb-b537-29b838225f7b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.490834 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf4xw\" (UniqueName: \"kubernetes.io/projected/9a13bb94-e9e5-40bb-b537-29b838225f7b-kube-api-access-gf4xw\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.490866 4936 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.490874 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.490884 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.490892 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.490901 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a13bb94-e9e5-40bb-b537-29b838225f7b-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.557861 4936 generic.go:334] "Generic (PLEG): container finished" podID="9a13bb94-e9e5-40bb-b537-29b838225f7b" containerID="8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726" exitCode=0 Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.557951 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" event={"ID":"9a13bb94-e9e5-40bb-b537-29b838225f7b","Type":"ContainerDied","Data":"8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726"} Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.557986 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" event={"ID":"9a13bb94-e9e5-40bb-b537-29b838225f7b","Type":"ContainerDied","Data":"a4007f9cdabf76a6e57b4009ac9d167043b9a77342f2ec00e580cc3232b39cf1"} Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.558007 4936 scope.go:117] "RemoveContainer" containerID="8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.558003 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-4tng4" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.564172 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerStarted","Data":"faad4399093da94d89f421818ff1d7b924e8da51f2e341b3f0de7d8e176ade08"} Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.567569 4936 generic.go:334] "Generic (PLEG): container finished" podID="b2a56dab-8614-4f51-bbe7-46f59dda60de" containerID="ca32faf9fcc7baa65af0e7c5ef5eb0a4a30b2e8d5d202bc7701f2b08d53b9ee9" exitCode=0 Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.567643 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" event={"ID":"b2a56dab-8614-4f51-bbe7-46f59dda60de","Type":"ContainerDied","Data":"ca32faf9fcc7baa65af0e7c5ef5eb0a4a30b2e8d5d202bc7701f2b08d53b9ee9"} Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.567661 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" event={"ID":"b2a56dab-8614-4f51-bbe7-46f59dda60de","Type":"ContainerStarted","Data":"4366334e4df96fce7c66687081cbd5d7d49b81f250b372b34a4f3a676ab5f807"} Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.568873 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.571205 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff274176-aa07-4573-b1f3-a6fa882b08b5","Type":"ContainerStarted","Data":"79d491e4c5b516ba232852f1300c49e51f2b7e50fa522dc82b520ae4682210fe"} Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.600883 4936 scope.go:117] "RemoveContainer" containerID="14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.612069 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" podStartSLOduration=3.612048199 podStartE2EDuration="3.612048199s" podCreationTimestamp="2026-03-20 16:20:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:10.590103598 +0000 UTC m=+1161.536471423" watchObservedRunningTime="2026-03-20 16:20:10.612048199 +0000 UTC m=+1161.558416014" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.640938 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-4tng4"] Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.651046 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-4tng4"] Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.699120 4936 scope.go:117] "RemoveContainer" containerID="8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726" Mar 20 16:20:10 crc kubenswrapper[4936]: E0320 16:20:10.699389 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726\": container with ID starting with 8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726 not found: ID does not exist" containerID="8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.699432 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726"} err="failed to get container status \"8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726\": rpc error: code = NotFound desc = could not find container \"8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726\": container with ID starting with 8af9b6a190bee6c76b5d399e1cce0c4ae4daae3e7936ceac3a142aef4bff8726 not found: ID does not exist" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.699464 4936 scope.go:117] "RemoveContainer" containerID="14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec" Mar 20 16:20:10 crc kubenswrapper[4936]: E0320 16:20:10.700313 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec\": container with ID starting with 14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec not found: ID does not exist" containerID="14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec" Mar 20 16:20:10 crc kubenswrapper[4936]: I0320 16:20:10.700344 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec"} err="failed to get container status \"14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec\": rpc error: code = NotFound desc = could not find container \"14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec\": container with ID starting with 14abf746bcb68a866c956afe8ca2e2a6db62634919c29e0af3d94e9350a771ec not found: ID does not exist" Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.588114 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerStarted","Data":"4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35"} Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.591584 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c","Type":"ContainerStarted","Data":"96436ca1fed35e10182e2607b33e5bc28ea66521bb8d329622a381e3b2c4ecef"} Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.591637 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c","Type":"ContainerStarted","Data":"7a74b36383a96d1236213256ac94b41fbf2d111ed4e16f6703371318953f0b49"} Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.593861 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerName="cinder-api-log" containerID="cri-o://79d491e4c5b516ba232852f1300c49e51f2b7e50fa522dc82b520ae4682210fe" gracePeriod=30 Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.593963 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerName="cinder-api" containerID="cri-o://867543d3c23aeeafc271619da6f94e519d190da920549ff81a7abbdf12c84a76" gracePeriod=30 Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.593802 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff274176-aa07-4573-b1f3-a6fa882b08b5","Type":"ContainerStarted","Data":"867543d3c23aeeafc271619da6f94e519d190da920549ff81a7abbdf12c84a76"} Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.594138 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.629979 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.455951884 podStartE2EDuration="4.629959734s" podCreationTimestamp="2026-03-20 16:20:07 +0000 UTC" firstStartedPulling="2026-03-20 16:20:08.579717532 +0000 UTC m=+1159.526085347" lastFinishedPulling="2026-03-20 16:20:09.753725392 +0000 UTC m=+1160.700093197" observedRunningTime="2026-03-20 16:20:11.619936681 +0000 UTC m=+1162.566304496" watchObservedRunningTime="2026-03-20 16:20:11.629959734 +0000 UTC m=+1162.576327549" Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.648179 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.648158455 podStartE2EDuration="4.648158455s" podCreationTimestamp="2026-03-20 16:20:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:11.642126209 +0000 UTC m=+1162.588494034" watchObservedRunningTime="2026-03-20 16:20:11.648158455 +0000 UTC m=+1162.594526270" Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.873261 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89ff8a26-ef18-4434-9601-11a47af62d20" path="/var/lib/kubelet/pods/89ff8a26-ef18-4434-9601-11a47af62d20/volumes" Mar 20 16:20:11 crc kubenswrapper[4936]: I0320 16:20:11.878916 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a13bb94-e9e5-40bb-b537-29b838225f7b" path="/var/lib/kubelet/pods/9a13bb94-e9e5-40bb-b537-29b838225f7b/volumes" Mar 20 16:20:12 crc kubenswrapper[4936]: I0320 16:20:12.213175 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:20:12 crc kubenswrapper[4936]: I0320 16:20:12.375474 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:20:12 crc kubenswrapper[4936]: I0320 16:20:12.649845 4936 generic.go:334] "Generic (PLEG): container finished" podID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerID="867543d3c23aeeafc271619da6f94e519d190da920549ff81a7abbdf12c84a76" exitCode=0 Mar 20 16:20:12 crc kubenswrapper[4936]: I0320 16:20:12.649891 4936 generic.go:334] "Generic (PLEG): container finished" podID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerID="79d491e4c5b516ba232852f1300c49e51f2b7e50fa522dc82b520ae4682210fe" exitCode=143 Mar 20 16:20:12 crc kubenswrapper[4936]: I0320 16:20:12.649962 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff274176-aa07-4573-b1f3-a6fa882b08b5","Type":"ContainerDied","Data":"867543d3c23aeeafc271619da6f94e519d190da920549ff81a7abbdf12c84a76"} Mar 20 16:20:12 crc kubenswrapper[4936]: I0320 16:20:12.649990 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff274176-aa07-4573-b1f3-a6fa882b08b5","Type":"ContainerDied","Data":"79d491e4c5b516ba232852f1300c49e51f2b7e50fa522dc82b520ae4682210fe"} Mar 20 16:20:12 crc kubenswrapper[4936]: I0320 16:20:12.671835 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerStarted","Data":"d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c"} Mar 20 16:20:12 crc kubenswrapper[4936]: I0320 16:20:12.971645 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.066731 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.258586 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data\") pod \"ff274176-aa07-4573-b1f3-a6fa882b08b5\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.260193 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data-custom\") pod \"ff274176-aa07-4573-b1f3-a6fa882b08b5\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.260229 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff274176-aa07-4573-b1f3-a6fa882b08b5-logs\") pod \"ff274176-aa07-4573-b1f3-a6fa882b08b5\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.260264 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djkk2\" (UniqueName: \"kubernetes.io/projected/ff274176-aa07-4573-b1f3-a6fa882b08b5-kube-api-access-djkk2\") pod \"ff274176-aa07-4573-b1f3-a6fa882b08b5\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.260314 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff274176-aa07-4573-b1f3-a6fa882b08b5-etc-machine-id\") pod \"ff274176-aa07-4573-b1f3-a6fa882b08b5\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.260340 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-scripts\") pod \"ff274176-aa07-4573-b1f3-a6fa882b08b5\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.260388 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-combined-ca-bundle\") pod \"ff274176-aa07-4573-b1f3-a6fa882b08b5\" (UID: \"ff274176-aa07-4573-b1f3-a6fa882b08b5\") " Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.260820 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ff274176-aa07-4573-b1f3-a6fa882b08b5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ff274176-aa07-4573-b1f3-a6fa882b08b5" (UID: "ff274176-aa07-4573-b1f3-a6fa882b08b5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.261787 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff274176-aa07-4573-b1f3-a6fa882b08b5-logs" (OuterVolumeSpecName: "logs") pod "ff274176-aa07-4573-b1f3-a6fa882b08b5" (UID: "ff274176-aa07-4573-b1f3-a6fa882b08b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.267634 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff274176-aa07-4573-b1f3-a6fa882b08b5-kube-api-access-djkk2" (OuterVolumeSpecName: "kube-api-access-djkk2") pod "ff274176-aa07-4573-b1f3-a6fa882b08b5" (UID: "ff274176-aa07-4573-b1f3-a6fa882b08b5"). InnerVolumeSpecName "kube-api-access-djkk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.267650 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ff274176-aa07-4573-b1f3-a6fa882b08b5" (UID: "ff274176-aa07-4573-b1f3-a6fa882b08b5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.280711 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-scripts" (OuterVolumeSpecName: "scripts") pod "ff274176-aa07-4573-b1f3-a6fa882b08b5" (UID: "ff274176-aa07-4573-b1f3-a6fa882b08b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.286144 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff274176-aa07-4573-b1f3-a6fa882b08b5" (UID: "ff274176-aa07-4573-b1f3-a6fa882b08b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.319734 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data" (OuterVolumeSpecName: "config-data") pod "ff274176-aa07-4573-b1f3-a6fa882b08b5" (UID: "ff274176-aa07-4573-b1f3-a6fa882b08b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.362804 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.362866 4936 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.362882 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff274176-aa07-4573-b1f3-a6fa882b08b5-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.362893 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djkk2\" (UniqueName: \"kubernetes.io/projected/ff274176-aa07-4573-b1f3-a6fa882b08b5-kube-api-access-djkk2\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.362907 4936 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff274176-aa07-4573-b1f3-a6fa882b08b5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.362917 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.362928 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff274176-aa07-4573-b1f3-a6fa882b08b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.681990 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff274176-aa07-4573-b1f3-a6fa882b08b5","Type":"ContainerDied","Data":"7d8fcb4839a638e8fe82b92955a7ec07897585cb29517cedf7fe76effb39e9e0"} Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.682033 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.682064 4936 scope.go:117] "RemoveContainer" containerID="867543d3c23aeeafc271619da6f94e519d190da920549ff81a7abbdf12c84a76" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.717831 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.726306 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.741147 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 20 16:20:13 crc kubenswrapper[4936]: E0320 16:20:13.741532 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a13bb94-e9e5-40bb-b537-29b838225f7b" containerName="dnsmasq-dns" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.742140 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a13bb94-e9e5-40bb-b537-29b838225f7b" containerName="dnsmasq-dns" Mar 20 16:20:13 crc kubenswrapper[4936]: E0320 16:20:13.742166 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerName="cinder-api-log" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.742174 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerName="cinder-api-log" Mar 20 16:20:13 crc kubenswrapper[4936]: E0320 16:20:13.742188 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05b79d13-f57b-455c-b5fe-b469c9bc0f45" containerName="oc" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.742197 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="05b79d13-f57b-455c-b5fe-b469c9bc0f45" containerName="oc" Mar 20 16:20:13 crc kubenswrapper[4936]: E0320 16:20:13.742213 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a13bb94-e9e5-40bb-b537-29b838225f7b" containerName="init" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.742220 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a13bb94-e9e5-40bb-b537-29b838225f7b" containerName="init" Mar 20 16:20:13 crc kubenswrapper[4936]: E0320 16:20:13.742236 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerName="cinder-api" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.742242 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerName="cinder-api" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.742400 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a13bb94-e9e5-40bb-b537-29b838225f7b" containerName="dnsmasq-dns" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.742415 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="05b79d13-f57b-455c-b5fe-b469c9bc0f45" containerName="oc" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.742430 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerName="cinder-api" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.742441 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff274176-aa07-4573-b1f3-a6fa882b08b5" containerName="cinder-api-log" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.746661 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.753291 4936 scope.go:117] "RemoveContainer" containerID="79d491e4c5b516ba232852f1300c49e51f2b7e50fa522dc82b520ae4682210fe" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.753605 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.754058 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.754215 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.781983 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.866132 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff274176-aa07-4573-b1f3-a6fa882b08b5" path="/var/lib/kubelet/pods/ff274176-aa07-4573-b1f3-a6fa882b08b5/volumes" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.874619 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-config-data\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.874734 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.874833 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k4kp\" (UniqueName: \"kubernetes.io/projected/2314f249-26ea-45c5-8225-ca99d8a4f82e-kube-api-access-6k4kp\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.874919 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.874948 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-scripts\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.874965 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-config-data-custom\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.875170 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2314f249-26ea-45c5-8225-ca99d8a4f82e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.875231 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2314f249-26ea-45c5-8225-ca99d8a4f82e-logs\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.875258 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.978397 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.978524 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k4kp\" (UniqueName: \"kubernetes.io/projected/2314f249-26ea-45c5-8225-ca99d8a4f82e-kube-api-access-6k4kp\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.978678 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.978716 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-scripts\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.978740 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-config-data-custom\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.978801 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2314f249-26ea-45c5-8225-ca99d8a4f82e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.978824 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2314f249-26ea-45c5-8225-ca99d8a4f82e-logs\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.978846 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.978897 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-config-data\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.979189 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2314f249-26ea-45c5-8225-ca99d8a4f82e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.980004 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2314f249-26ea-45c5-8225-ca99d8a4f82e-logs\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.988149 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-scripts\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.990080 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.991175 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-config-data-custom\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.993085 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.996158 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:13 crc kubenswrapper[4936]: I0320 16:20:13.996948 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2314f249-26ea-45c5-8225-ca99d8a4f82e-config-data\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.008891 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k4kp\" (UniqueName: \"kubernetes.io/projected/2314f249-26ea-45c5-8225-ca99d8a4f82e-kube-api-access-6k4kp\") pod \"cinder-api-0\" (UID: \"2314f249-26ea-45c5-8225-ca99d8a4f82e\") " pod="openstack/cinder-api-0" Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.076275 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.168296 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.174764 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6ff4f57d48-qp24t" Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.285251 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79458c9dc6-vlzgk"] Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.538267 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.591136 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 20 16:20:14 crc kubenswrapper[4936]: W0320 16:20:14.599125 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2314f249_26ea_45c5_8225_ca99d8a4f82e.slice/crio-a87b4ed13ba25a8433465dab2a47816c7c3af6cfb007f95b9cae5e9526629d23 WatchSource:0}: Error finding container a87b4ed13ba25a8433465dab2a47816c7c3af6cfb007f95b9cae5e9526629d23: Status 404 returned error can't find the container with id a87b4ed13ba25a8433465dab2a47816c7c3af6cfb007f95b9cae5e9526629d23 Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.675904 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.748090 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerStarted","Data":"b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9"} Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.749229 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2314f249-26ea-45c5-8225-ca99d8a4f82e","Type":"ContainerStarted","Data":"a87b4ed13ba25a8433465dab2a47816c7c3af6cfb007f95b9cae5e9526629d23"} Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.750110 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79458c9dc6-vlzgk" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon-log" containerID="cri-o://ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4" gracePeriod=30 Mar 20 16:20:14 crc kubenswrapper[4936]: I0320 16:20:14.751729 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79458c9dc6-vlzgk" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon" containerID="cri-o://b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9" gracePeriod=30 Mar 20 16:20:15 crc kubenswrapper[4936]: I0320 16:20:15.772624 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2314f249-26ea-45c5-8225-ca99d8a4f82e","Type":"ContainerStarted","Data":"57090dd2e173486a06ba99fdf3cec2467a5ec3933fbab380097b1490843f183d"} Mar 20 16:20:16 crc kubenswrapper[4936]: I0320 16:20:16.786496 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2314f249-26ea-45c5-8225-ca99d8a4f82e","Type":"ContainerStarted","Data":"0dd8281d4924d974eae7575d95eca1358204ad8db0fb8aa4656bf5e316563bdd"} Mar 20 16:20:16 crc kubenswrapper[4936]: I0320 16:20:16.788592 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 20 16:20:16 crc kubenswrapper[4936]: I0320 16:20:16.825713 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.825693085 podStartE2EDuration="3.825693085s" podCreationTimestamp="2026-03-20 16:20:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:16.818587058 +0000 UTC m=+1167.764954883" watchObservedRunningTime="2026-03-20 16:20:16.825693085 +0000 UTC m=+1167.772060900" Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.211310 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.213999 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-598cbc676b-tm5b2" Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.285835 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c9dff68b4-6qmrf"] Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.286099 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c9dff68b4-6qmrf" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api-log" containerID="cri-o://cddaadd66f02ccb8e7888193cf6b62fe6fc34d4566ea8e87c4a4fbd5b3c5d6a9" gracePeriod=30 Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.286188 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c9dff68b4-6qmrf" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api" containerID="cri-o://656380c7600e02c0c0e14072bc9a0e44503c8ea8975917d7d4d207dec2f80f4f" gracePeriod=30 Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.803862 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerStarted","Data":"001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522"} Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.804226 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.805604 4936 generic.go:334] "Generic (PLEG): container finished" podID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerID="cddaadd66f02ccb8e7888193cf6b62fe6fc34d4566ea8e87c4a4fbd5b3c5d6a9" exitCode=143 Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.806002 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c9dff68b4-6qmrf" event={"ID":"802be0f0-c896-47de-8efa-24c499ed3e2c","Type":"ContainerDied","Data":"cddaadd66f02ccb8e7888193cf6b62fe6fc34d4566ea8e87c4a4fbd5b3c5d6a9"} Mar 20 16:20:17 crc kubenswrapper[4936]: I0320 16:20:17.840605 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.669050802 podStartE2EDuration="9.840584101s" podCreationTimestamp="2026-03-20 16:20:08 +0000 UTC" firstStartedPulling="2026-03-20 16:20:09.665276881 +0000 UTC m=+1160.611644696" lastFinishedPulling="2026-03-20 16:20:16.83681018 +0000 UTC m=+1167.783177995" observedRunningTime="2026-03-20 16:20:17.824352637 +0000 UTC m=+1168.770720462" watchObservedRunningTime="2026-03-20 16:20:17.840584101 +0000 UTC m=+1168.786951916" Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.195029 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.266261 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.293709 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.382073 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-s95m6"] Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.382340 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" podUID="70c0f7b6-fa50-4b4a-9291-596d0b403576" containerName="dnsmasq-dns" containerID="cri-o://81d0c8cc8be8bd2d4588928bd77ac428b077b959169acc5b8bbdbdc9264380d8" gracePeriod=10 Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.821707 4936 generic.go:334] "Generic (PLEG): container finished" podID="70c0f7b6-fa50-4b4a-9291-596d0b403576" containerID="81d0c8cc8be8bd2d4588928bd77ac428b077b959169acc5b8bbdbdc9264380d8" exitCode=0 Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.821790 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" event={"ID":"70c0f7b6-fa50-4b4a-9291-596d0b403576","Type":"ContainerDied","Data":"81d0c8cc8be8bd2d4588928bd77ac428b077b959169acc5b8bbdbdc9264380d8"} Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.825115 4936 generic.go:334] "Generic (PLEG): container finished" podID="02669555-c7e4-461a-86bd-aa30db944ab9" containerID="b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9" exitCode=0 Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.825205 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79458c9dc6-vlzgk" event={"ID":"02669555-c7e4-461a-86bd-aa30db944ab9","Type":"ContainerDied","Data":"b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9"} Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.825641 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerName="cinder-scheduler" containerID="cri-o://7a74b36383a96d1236213256ac94b41fbf2d111ed4e16f6703371318953f0b49" gracePeriod=30 Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.825834 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerName="probe" containerID="cri-o://96436ca1fed35e10182e2607b33e5bc28ea66521bb8d329622a381e3b2c4ecef" gracePeriod=30 Mar 20 16:20:18 crc kubenswrapper[4936]: I0320 16:20:18.975216 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.103206 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-svc\") pod \"70c0f7b6-fa50-4b4a-9291-596d0b403576\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.103286 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-swift-storage-0\") pod \"70c0f7b6-fa50-4b4a-9291-596d0b403576\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.103341 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-nb\") pod \"70c0f7b6-fa50-4b4a-9291-596d0b403576\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.103403 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x5lp\" (UniqueName: \"kubernetes.io/projected/70c0f7b6-fa50-4b4a-9291-596d0b403576-kube-api-access-4x5lp\") pod \"70c0f7b6-fa50-4b4a-9291-596d0b403576\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.103477 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-config\") pod \"70c0f7b6-fa50-4b4a-9291-596d0b403576\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.103583 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-sb\") pod \"70c0f7b6-fa50-4b4a-9291-596d0b403576\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.111908 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70c0f7b6-fa50-4b4a-9291-596d0b403576-kube-api-access-4x5lp" (OuterVolumeSpecName: "kube-api-access-4x5lp") pod "70c0f7b6-fa50-4b4a-9291-596d0b403576" (UID: "70c0f7b6-fa50-4b4a-9291-596d0b403576"). InnerVolumeSpecName "kube-api-access-4x5lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.151631 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "70c0f7b6-fa50-4b4a-9291-596d0b403576" (UID: "70c0f7b6-fa50-4b4a-9291-596d0b403576"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.154160 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-config" (OuterVolumeSpecName: "config") pod "70c0f7b6-fa50-4b4a-9291-596d0b403576" (UID: "70c0f7b6-fa50-4b4a-9291-596d0b403576"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.158423 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "70c0f7b6-fa50-4b4a-9291-596d0b403576" (UID: "70c0f7b6-fa50-4b4a-9291-596d0b403576"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:19 crc kubenswrapper[4936]: E0320 16:20:19.158850 4936 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-swift-storage-0 podName:70c0f7b6-fa50-4b4a-9291-596d0b403576 nodeName:}" failed. No retries permitted until 2026-03-20 16:20:19.658816059 +0000 UTC m=+1170.605183944 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-swift-storage-0" (UniqueName: "kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-swift-storage-0") pod "70c0f7b6-fa50-4b4a-9291-596d0b403576" (UID: "70c0f7b6-fa50-4b4a-9291-596d0b403576") : error deleting /var/lib/kubelet/pods/70c0f7b6-fa50-4b4a-9291-596d0b403576/volume-subpaths: remove /var/lib/kubelet/pods/70c0f7b6-fa50-4b4a-9291-596d0b403576/volume-subpaths: no such file or directory Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.159262 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "70c0f7b6-fa50-4b4a-9291-596d0b403576" (UID: "70c0f7b6-fa50-4b4a-9291-596d0b403576"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.210594 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x5lp\" (UniqueName: \"kubernetes.io/projected/70c0f7b6-fa50-4b4a-9291-596d0b403576-kube-api-access-4x5lp\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.210633 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.210645 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.210656 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.210668 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.718395 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-swift-storage-0\") pod \"70c0f7b6-fa50-4b4a-9291-596d0b403576\" (UID: \"70c0f7b6-fa50-4b4a-9291-596d0b403576\") " Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.719020 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "70c0f7b6-fa50-4b4a-9291-596d0b403576" (UID: "70c0f7b6-fa50-4b4a-9291-596d0b403576"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.719664 4936 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70c0f7b6-fa50-4b4a-9291-596d0b403576-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.842473 4936 generic.go:334] "Generic (PLEG): container finished" podID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerID="96436ca1fed35e10182e2607b33e5bc28ea66521bb8d329622a381e3b2c4ecef" exitCode=0 Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.842589 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c","Type":"ContainerDied","Data":"96436ca1fed35e10182e2607b33e5bc28ea66521bb8d329622a381e3b2c4ecef"} Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.846001 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" event={"ID":"70c0f7b6-fa50-4b4a-9291-596d0b403576","Type":"ContainerDied","Data":"a3d333ccdb3ffa3bfd1151e1da820b55d65909d5c66aff195484d7c573c11856"} Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.846056 4936 scope.go:117] "RemoveContainer" containerID="81d0c8cc8be8bd2d4588928bd77ac428b077b959169acc5b8bbdbdc9264380d8" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.846168 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-s95m6" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.850197 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79458c9dc6-vlzgk" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.874044 4936 scope.go:117] "RemoveContainer" containerID="57126127a3e4442f23fa9adde47e5ec10b626fe50b533569e9e19edd1a95cda8" Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.886745 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-s95m6"] Mar 20 16:20:19 crc kubenswrapper[4936]: I0320 16:20:19.897102 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-s95m6"] Mar 20 16:20:20 crc kubenswrapper[4936]: I0320 16:20:20.435454 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c9dff68b4-6qmrf" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:58468->10.217.0.165:9311: read: connection reset by peer" Mar 20 16:20:20 crc kubenswrapper[4936]: I0320 16:20:20.435762 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c9dff68b4-6qmrf" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:58466->10.217.0.165:9311: read: connection reset by peer" Mar 20 16:20:20 crc kubenswrapper[4936]: I0320 16:20:20.861573 4936 generic.go:334] "Generic (PLEG): container finished" podID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerID="656380c7600e02c0c0e14072bc9a0e44503c8ea8975917d7d4d207dec2f80f4f" exitCode=0 Mar 20 16:20:20 crc kubenswrapper[4936]: I0320 16:20:20.861675 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c9dff68b4-6qmrf" event={"ID":"802be0f0-c896-47de-8efa-24c499ed3e2c","Type":"ContainerDied","Data":"656380c7600e02c0c0e14072bc9a0e44503c8ea8975917d7d4d207dec2f80f4f"} Mar 20 16:20:20 crc kubenswrapper[4936]: I0320 16:20:20.968418 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.049387 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-combined-ca-bundle\") pod \"802be0f0-c896-47de-8efa-24c499ed3e2c\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.049443 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802be0f0-c896-47de-8efa-24c499ed3e2c-logs\") pod \"802be0f0-c896-47de-8efa-24c499ed3e2c\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.049469 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data-custom\") pod \"802be0f0-c896-47de-8efa-24c499ed3e2c\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.049557 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccjj4\" (UniqueName: \"kubernetes.io/projected/802be0f0-c896-47de-8efa-24c499ed3e2c-kube-api-access-ccjj4\") pod \"802be0f0-c896-47de-8efa-24c499ed3e2c\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.049575 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data\") pod \"802be0f0-c896-47de-8efa-24c499ed3e2c\" (UID: \"802be0f0-c896-47de-8efa-24c499ed3e2c\") " Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.049921 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/802be0f0-c896-47de-8efa-24c499ed3e2c-logs" (OuterVolumeSpecName: "logs") pod "802be0f0-c896-47de-8efa-24c499ed3e2c" (UID: "802be0f0-c896-47de-8efa-24c499ed3e2c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.067796 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "802be0f0-c896-47de-8efa-24c499ed3e2c" (UID: "802be0f0-c896-47de-8efa-24c499ed3e2c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.067891 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/802be0f0-c896-47de-8efa-24c499ed3e2c-kube-api-access-ccjj4" (OuterVolumeSpecName: "kube-api-access-ccjj4") pod "802be0f0-c896-47de-8efa-24c499ed3e2c" (UID: "802be0f0-c896-47de-8efa-24c499ed3e2c"). InnerVolumeSpecName "kube-api-access-ccjj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.080761 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "802be0f0-c896-47de-8efa-24c499ed3e2c" (UID: "802be0f0-c896-47de-8efa-24c499ed3e2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.099741 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data" (OuterVolumeSpecName: "config-data") pod "802be0f0-c896-47de-8efa-24c499ed3e2c" (UID: "802be0f0-c896-47de-8efa-24c499ed3e2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.151921 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.151949 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802be0f0-c896-47de-8efa-24c499ed3e2c-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.151957 4936 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.151969 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccjj4\" (UniqueName: \"kubernetes.io/projected/802be0f0-c896-47de-8efa-24c499ed3e2c-kube-api-access-ccjj4\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.151977 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802be0f0-c896-47de-8efa-24c499ed3e2c-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.461001 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.873418 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70c0f7b6-fa50-4b4a-9291-596d0b403576" path="/var/lib/kubelet/pods/70c0f7b6-fa50-4b4a-9291-596d0b403576/volumes" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.887004 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c9dff68b4-6qmrf" event={"ID":"802be0f0-c896-47de-8efa-24c499ed3e2c","Type":"ContainerDied","Data":"45ee1d9e85d6473da389796355687f44adc633538e08ce4f344dd54321d7c785"} Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.887068 4936 scope.go:117] "RemoveContainer" containerID="656380c7600e02c0c0e14072bc9a0e44503c8ea8975917d7d4d207dec2f80f4f" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.887322 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c9dff68b4-6qmrf" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.924588 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c9dff68b4-6qmrf"] Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.930484 4936 scope.go:117] "RemoveContainer" containerID="cddaadd66f02ccb8e7888193cf6b62fe6fc34d4566ea8e87c4a4fbd5b3c5d6a9" Mar 20 16:20:21 crc kubenswrapper[4936]: I0320 16:20:21.931310 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7c9dff68b4-6qmrf"] Mar 20 16:20:22 crc kubenswrapper[4936]: I0320 16:20:22.668885 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-768674d4fd-4vgnw" Mar 20 16:20:22 crc kubenswrapper[4936]: I0320 16:20:22.909238 4936 generic.go:334] "Generic (PLEG): container finished" podID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerID="7a74b36383a96d1236213256ac94b41fbf2d111ed4e16f6703371318953f0b49" exitCode=0 Mar 20 16:20:22 crc kubenswrapper[4936]: I0320 16:20:22.909306 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c","Type":"ContainerDied","Data":"7a74b36383a96d1236213256ac94b41fbf2d111ed4e16f6703371318953f0b49"} Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.175358 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.197617 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzdst\" (UniqueName: \"kubernetes.io/projected/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-kube-api-access-nzdst\") pod \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.197658 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-etc-machine-id\") pod \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.197706 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-combined-ca-bundle\") pod \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.197740 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data-custom\") pod \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.197772 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data\") pod \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.197793 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" (UID: "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.197917 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-scripts\") pod \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\" (UID: \"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c\") " Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.198221 4936 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.203670 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-kube-api-access-nzdst" (OuterVolumeSpecName: "kube-api-access-nzdst") pod "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" (UID: "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c"). InnerVolumeSpecName "kube-api-access-nzdst". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.206651 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-scripts" (OuterVolumeSpecName: "scripts") pod "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" (UID: "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.208391 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" (UID: "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.265377 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" (UID: "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.299900 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.299933 4936 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.299946 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.299983 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzdst\" (UniqueName: \"kubernetes.io/projected/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-kube-api-access-nzdst\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.331206 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data" (OuterVolumeSpecName: "config-data") pod "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" (UID: "1f09bb09-5fa2-41d9-8eeb-676d3ee6616c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.402691 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.866898 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" path="/var/lib/kubelet/pods/802be0f0-c896-47de-8efa-24c499ed3e2c/volumes" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.920994 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1f09bb09-5fa2-41d9-8eeb-676d3ee6616c","Type":"ContainerDied","Data":"46a76fee5b45fa1fb21d40c15d111803c7a5b5057f2ac9c1fb794da93e1caddc"} Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.921271 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.922075 4936 scope.go:117] "RemoveContainer" containerID="96436ca1fed35e10182e2607b33e5bc28ea66521bb8d329622a381e3b2c4ecef" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.950187 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.953820 4936 scope.go:117] "RemoveContainer" containerID="7a74b36383a96d1236213256ac94b41fbf2d111ed4e16f6703371318953f0b49" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.959346 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981141 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 20 16:20:23 crc kubenswrapper[4936]: E0320 16:20:23.981588 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api-log" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981611 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api-log" Mar 20 16:20:23 crc kubenswrapper[4936]: E0320 16:20:23.981633 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerName="cinder-scheduler" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981643 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerName="cinder-scheduler" Mar 20 16:20:23 crc kubenswrapper[4936]: E0320 16:20:23.981664 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c0f7b6-fa50-4b4a-9291-596d0b403576" containerName="dnsmasq-dns" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981673 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c0f7b6-fa50-4b4a-9291-596d0b403576" containerName="dnsmasq-dns" Mar 20 16:20:23 crc kubenswrapper[4936]: E0320 16:20:23.981685 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerName="probe" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981693 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerName="probe" Mar 20 16:20:23 crc kubenswrapper[4936]: E0320 16:20:23.981715 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981723 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api" Mar 20 16:20:23 crc kubenswrapper[4936]: E0320 16:20:23.981735 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c0f7b6-fa50-4b4a-9291-596d0b403576" containerName="init" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981743 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c0f7b6-fa50-4b4a-9291-596d0b403576" containerName="init" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981944 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api-log" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981960 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="802be0f0-c896-47de-8efa-24c499ed3e2c" containerName="barbican-api" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.981980 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c0f7b6-fa50-4b4a-9291-596d0b403576" containerName="dnsmasq-dns" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.982000 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerName="probe" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.982015 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" containerName="cinder-scheduler" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.983171 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.985151 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 20 16:20:23 crc kubenswrapper[4936]: I0320 16:20:23.996000 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.014170 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.014304 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.014393 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.014421 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr6zd\" (UniqueName: \"kubernetes.io/projected/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-kube-api-access-gr6zd\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.014452 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-config-data\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.014475 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-scripts\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.116151 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.116592 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.116626 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr6zd\" (UniqueName: \"kubernetes.io/projected/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-kube-api-access-gr6zd\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.116676 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-config-data\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.116705 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-scripts\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.116766 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.116853 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.120311 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.120400 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-scripts\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.122637 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-config-data\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.123413 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.136556 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr6zd\" (UniqueName: \"kubernetes.io/projected/46cbf2a6-9c75-49a6-b7b1-31fcf600d874-kube-api-access-gr6zd\") pod \"cinder-scheduler-0\" (UID: \"46cbf2a6-9c75-49a6-b7b1-31fcf600d874\") " pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.204310 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-698b9689fc-cph26" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.277578 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c66f596d-qcxkb"] Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.277876 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c66f596d-qcxkb" podUID="7da58715-94c1-487a-9f20-f008b4bea305" containerName="neutron-httpd" containerID="cri-o://5708ae6c1402aa64eda9c419b36f83f475ee4a5b700085736fe1fb6898b678fa" gracePeriod=30 Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.277839 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c66f596d-qcxkb" podUID="7da58715-94c1-487a-9f20-f008b4bea305" containerName="neutron-api" containerID="cri-o://4cd46a430f3c3a8c5a6186f91c234b96eae73ad4dbe2e65fc9287325d41fe261" gracePeriod=30 Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.352749 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.883482 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.944742 4936 generic.go:334] "Generic (PLEG): container finished" podID="7da58715-94c1-487a-9f20-f008b4bea305" containerID="5708ae6c1402aa64eda9c419b36f83f475ee4a5b700085736fe1fb6898b678fa" exitCode=0 Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.944831 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c66f596d-qcxkb" event={"ID":"7da58715-94c1-487a-9f20-f008b4bea305","Type":"ContainerDied","Data":"5708ae6c1402aa64eda9c419b36f83f475ee4a5b700085736fe1fb6898b678fa"} Mar 20 16:20:24 crc kubenswrapper[4936]: I0320 16:20:24.955200 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46cbf2a6-9c75-49a6-b7b1-31fcf600d874","Type":"ContainerStarted","Data":"610218fbe2bec17cc9f266553739869f3cfe02a210e15007be93ebece5a5e317"} Mar 20 16:20:25 crc kubenswrapper[4936]: I0320 16:20:25.869729 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f09bb09-5fa2-41d9-8eeb-676d3ee6616c" path="/var/lib/kubelet/pods/1f09bb09-5fa2-41d9-8eeb-676d3ee6616c/volumes" Mar 20 16:20:25 crc kubenswrapper[4936]: I0320 16:20:25.974519 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46cbf2a6-9c75-49a6-b7b1-31fcf600d874","Type":"ContainerStarted","Data":"cd86dec759b546c034e832292f4464b86484faa718d6a28f45999e1b52b3cde2"} Mar 20 16:20:26 crc kubenswrapper[4936]: I0320 16:20:26.278340 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 20 16:20:26 crc kubenswrapper[4936]: I0320 16:20:26.987585 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46cbf2a6-9c75-49a6-b7b1-31fcf600d874","Type":"ContainerStarted","Data":"90ca2f6359ede36254c6dd3eae3e14cfb548aff4f34f0a01481d74919c564a88"} Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.017982 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-74b5cc8579-wqhwv"] Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.019639 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.022425 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.022648 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.023580 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.032464 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.032443318 podStartE2EDuration="4.032443318s" podCreationTimestamp="2026-03-20 16:20:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:27.008004111 +0000 UTC m=+1177.954371926" watchObservedRunningTime="2026-03-20 16:20:27.032443318 +0000 UTC m=+1177.978811133" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.045519 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-74b5cc8579-wqhwv"] Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.180775 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9284dab5-bdd1-4792-8944-0f9c820a927d-etc-swift\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.181170 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2clr\" (UniqueName: \"kubernetes.io/projected/9284dab5-bdd1-4792-8944-0f9c820a927d-kube-api-access-j2clr\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.181323 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-internal-tls-certs\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.181408 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-public-tls-certs\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.181448 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-config-data\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.181478 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-combined-ca-bundle\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.181509 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9284dab5-bdd1-4792-8944-0f9c820a927d-log-httpd\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.181921 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9284dab5-bdd1-4792-8944-0f9c820a927d-run-httpd\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.283492 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-internal-tls-certs\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.283553 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-public-tls-certs\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.283580 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-config-data\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.283605 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-combined-ca-bundle\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.283630 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9284dab5-bdd1-4792-8944-0f9c820a927d-log-httpd\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.283689 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9284dab5-bdd1-4792-8944-0f9c820a927d-run-httpd\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.283711 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9284dab5-bdd1-4792-8944-0f9c820a927d-etc-swift\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.283736 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2clr\" (UniqueName: \"kubernetes.io/projected/9284dab5-bdd1-4792-8944-0f9c820a927d-kube-api-access-j2clr\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.284367 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9284dab5-bdd1-4792-8944-0f9c820a927d-log-httpd\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.284654 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9284dab5-bdd1-4792-8944-0f9c820a927d-run-httpd\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.289859 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-combined-ca-bundle\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.292073 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-internal-tls-certs\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.292567 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-config-data\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.293837 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9284dab5-bdd1-4792-8944-0f9c820a927d-etc-swift\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.297696 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9284dab5-bdd1-4792-8944-0f9c820a927d-public-tls-certs\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.310201 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2clr\" (UniqueName: \"kubernetes.io/projected/9284dab5-bdd1-4792-8944-0f9c820a927d-kube-api-access-j2clr\") pod \"swift-proxy-74b5cc8579-wqhwv\" (UID: \"9284dab5-bdd1-4792-8944-0f9c820a927d\") " pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.335948 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.568137 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.569740 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.572563 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.572917 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.574298 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-bpvmj" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.584670 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.594013 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/961ea539-f955-46fd-8afa-8ec5b42133eb-openstack-config-secret\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.594079 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961ea539-f955-46fd-8afa-8ec5b42133eb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.594123 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwvdq\" (UniqueName: \"kubernetes.io/projected/961ea539-f955-46fd-8afa-8ec5b42133eb-kube-api-access-xwvdq\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.594258 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/961ea539-f955-46fd-8afa-8ec5b42133eb-openstack-config\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.696744 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/961ea539-f955-46fd-8afa-8ec5b42133eb-openstack-config-secret\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.696831 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961ea539-f955-46fd-8afa-8ec5b42133eb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.696878 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwvdq\" (UniqueName: \"kubernetes.io/projected/961ea539-f955-46fd-8afa-8ec5b42133eb-kube-api-access-xwvdq\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.697029 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/961ea539-f955-46fd-8afa-8ec5b42133eb-openstack-config\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.698137 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/961ea539-f955-46fd-8afa-8ec5b42133eb-openstack-config\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.712246 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961ea539-f955-46fd-8afa-8ec5b42133eb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.714583 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/961ea539-f955-46fd-8afa-8ec5b42133eb-openstack-config-secret\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.724126 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwvdq\" (UniqueName: \"kubernetes.io/projected/961ea539-f955-46fd-8afa-8ec5b42133eb-kube-api-access-xwvdq\") pod \"openstackclient\" (UID: \"961ea539-f955-46fd-8afa-8ec5b42133eb\") " pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.940270 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 20 16:20:27 crc kubenswrapper[4936]: I0320 16:20:27.997583 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-74b5cc8579-wqhwv"] Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.019266 4936 generic.go:334] "Generic (PLEG): container finished" podID="7da58715-94c1-487a-9f20-f008b4bea305" containerID="4cd46a430f3c3a8c5a6186f91c234b96eae73ad4dbe2e65fc9287325d41fe261" exitCode=0 Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.019362 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c66f596d-qcxkb" event={"ID":"7da58715-94c1-487a-9f20-f008b4bea305","Type":"ContainerDied","Data":"4cd46a430f3c3a8c5a6186f91c234b96eae73ad4dbe2e65fc9287325d41fe261"} Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.426217 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.585355 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.609877 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjjsh\" (UniqueName: \"kubernetes.io/projected/7da58715-94c1-487a-9f20-f008b4bea305-kube-api-access-zjjsh\") pod \"7da58715-94c1-487a-9f20-f008b4bea305\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.610087 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-httpd-config\") pod \"7da58715-94c1-487a-9f20-f008b4bea305\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.610137 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-config\") pod \"7da58715-94c1-487a-9f20-f008b4bea305\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.610270 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-combined-ca-bundle\") pod \"7da58715-94c1-487a-9f20-f008b4bea305\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.610354 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-ovndb-tls-certs\") pod \"7da58715-94c1-487a-9f20-f008b4bea305\" (UID: \"7da58715-94c1-487a-9f20-f008b4bea305\") " Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.625915 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "7da58715-94c1-487a-9f20-f008b4bea305" (UID: "7da58715-94c1-487a-9f20-f008b4bea305"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.637782 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7da58715-94c1-487a-9f20-f008b4bea305-kube-api-access-zjjsh" (OuterVolumeSpecName: "kube-api-access-zjjsh") pod "7da58715-94c1-487a-9f20-f008b4bea305" (UID: "7da58715-94c1-487a-9f20-f008b4bea305"). InnerVolumeSpecName "kube-api-access-zjjsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.672182 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-config" (OuterVolumeSpecName: "config") pod "7da58715-94c1-487a-9f20-f008b4bea305" (UID: "7da58715-94c1-487a-9f20-f008b4bea305"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.689630 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7da58715-94c1-487a-9f20-f008b4bea305" (UID: "7da58715-94c1-487a-9f20-f008b4bea305"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.711867 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.711914 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjjsh\" (UniqueName: \"kubernetes.io/projected/7da58715-94c1-487a-9f20-f008b4bea305-kube-api-access-zjjsh\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.711929 4936 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.711940 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.723400 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "7da58715-94c1-487a-9f20-f008b4bea305" (UID: "7da58715-94c1-487a-9f20-f008b4bea305"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:28 crc kubenswrapper[4936]: I0320 16:20:28.814026 4936 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da58715-94c1-487a-9f20-f008b4bea305-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.029210 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"961ea539-f955-46fd-8afa-8ec5b42133eb","Type":"ContainerStarted","Data":"ad6c8707310df0e61dcaea9ad07fd637af8edf62c0179e2c718694b6e87acad8"} Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.032256 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c66f596d-qcxkb" Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.032252 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c66f596d-qcxkb" event={"ID":"7da58715-94c1-487a-9f20-f008b4bea305","Type":"ContainerDied","Data":"bc91e8a47866e95e60f1cd4efb677f88684ce2b83cace52de98a108e74f91c3f"} Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.032398 4936 scope.go:117] "RemoveContainer" containerID="5708ae6c1402aa64eda9c419b36f83f475ee4a5b700085736fe1fb6898b678fa" Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.034958 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74b5cc8579-wqhwv" event={"ID":"9284dab5-bdd1-4792-8944-0f9c820a927d","Type":"ContainerStarted","Data":"6e8ea8d278f0dbc60d0ca9e6a5b9f1460c4b0cb6e9a4996e37ffab3da9940310"} Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.034985 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74b5cc8579-wqhwv" event={"ID":"9284dab5-bdd1-4792-8944-0f9c820a927d","Type":"ContainerStarted","Data":"954f8a822ab579b23f78948e52c9fc9fa5a0dce06d3f617c5e8d96c6cc583c0b"} Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.034997 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74b5cc8579-wqhwv" event={"ID":"9284dab5-bdd1-4792-8944-0f9c820a927d","Type":"ContainerStarted","Data":"e985843d22b591f23ace31e69fa7d5852574702281420b209189c269bcb4fc9b"} Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.035165 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.062386 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-74b5cc8579-wqhwv" podStartSLOduration=3.062370843 podStartE2EDuration="3.062370843s" podCreationTimestamp="2026-03-20 16:20:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:29.060891092 +0000 UTC m=+1180.007258907" watchObservedRunningTime="2026-03-20 16:20:29.062370843 +0000 UTC m=+1180.008738658" Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.068845 4936 scope.go:117] "RemoveContainer" containerID="4cd46a430f3c3a8c5a6186f91c234b96eae73ad4dbe2e65fc9287325d41fe261" Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.101607 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c66f596d-qcxkb"] Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.114321 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c66f596d-qcxkb"] Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.339651 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.339913 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="ceilometer-central-agent" containerID="cri-o://4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35" gracePeriod=30 Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.340503 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="proxy-httpd" containerID="cri-o://001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522" gracePeriod=30 Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.340743 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="ceilometer-notification-agent" containerID="cri-o://d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c" gracePeriod=30 Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.340796 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="sg-core" containerID="cri-o://b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9" gracePeriod=30 Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.350880 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.353620 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.850054 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79458c9dc6-vlzgk" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 20 16:20:29 crc kubenswrapper[4936]: I0320 16:20:29.864132 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7da58715-94c1-487a-9f20-f008b4bea305" path="/var/lib/kubelet/pods/7da58715-94c1-487a-9f20-f008b4bea305/volumes" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.047094 4936 generic.go:334] "Generic (PLEG): container finished" podID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerID="001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522" exitCode=0 Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.047130 4936 generic.go:334] "Generic (PLEG): container finished" podID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerID="b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9" exitCode=2 Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.047141 4936 generic.go:334] "Generic (PLEG): container finished" podID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerID="d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c" exitCode=0 Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.047198 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerDied","Data":"001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522"} Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.047247 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerDied","Data":"b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9"} Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.047263 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerDied","Data":"d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c"} Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.049382 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.651425 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.664014 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b4br\" (UniqueName: \"kubernetes.io/projected/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-kube-api-access-2b4br\") pod \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.664476 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-log-httpd\") pod \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.664600 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-combined-ca-bundle\") pod \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.664726 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-scripts\") pod \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.664826 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-config-data\") pod \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.664968 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-run-httpd\") pod \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.665079 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-sg-core-conf-yaml\") pod \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\" (UID: \"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c\") " Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.666996 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" (UID: "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.667025 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" (UID: "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.707948 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-scripts" (OuterVolumeSpecName: "scripts") pod "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" (UID: "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.711972 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-kube-api-access-2b4br" (OuterVolumeSpecName: "kube-api-access-2b4br") pod "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" (UID: "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c"). InnerVolumeSpecName "kube-api-access-2b4br". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.768079 4936 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.768260 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.768350 4936 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.768405 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b4br\" (UniqueName: \"kubernetes.io/projected/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-kube-api-access-2b4br\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.823700 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" (UID: "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.869894 4936 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.948841 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-config-data" (OuterVolumeSpecName: "config-data") pod "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" (UID: "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.967663 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" (UID: "d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.971861 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:30 crc kubenswrapper[4936]: I0320 16:20:30.971889 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.068737 4936 generic.go:334] "Generic (PLEG): container finished" podID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerID="4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35" exitCode=0 Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.068784 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerDied","Data":"4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35"} Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.069021 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c","Type":"ContainerDied","Data":"faad4399093da94d89f421818ff1d7b924e8da51f2e341b3f0de7d8e176ade08"} Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.069057 4936 scope.go:117] "RemoveContainer" containerID="001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.068846 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.105386 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.122535 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.123205 4936 scope.go:117] "RemoveContainer" containerID="b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.149520 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:31 crc kubenswrapper[4936]: E0320 16:20:31.150511 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="proxy-httpd" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.150735 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="proxy-httpd" Mar 20 16:20:31 crc kubenswrapper[4936]: E0320 16:20:31.150763 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="ceilometer-central-agent" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.150777 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="ceilometer-central-agent" Mar 20 16:20:31 crc kubenswrapper[4936]: E0320 16:20:31.151207 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="ceilometer-notification-agent" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.151228 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="ceilometer-notification-agent" Mar 20 16:20:31 crc kubenswrapper[4936]: E0320 16:20:31.151243 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da58715-94c1-487a-9f20-f008b4bea305" containerName="neutron-httpd" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.151252 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da58715-94c1-487a-9f20-f008b4bea305" containerName="neutron-httpd" Mar 20 16:20:31 crc kubenswrapper[4936]: E0320 16:20:31.151275 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="sg-core" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.151692 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="sg-core" Mar 20 16:20:31 crc kubenswrapper[4936]: E0320 16:20:31.151710 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da58715-94c1-487a-9f20-f008b4bea305" containerName="neutron-api" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.151719 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da58715-94c1-487a-9f20-f008b4bea305" containerName="neutron-api" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.151941 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="ceilometer-central-agent" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.151961 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="7da58715-94c1-487a-9f20-f008b4bea305" containerName="neutron-api" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.151976 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="ceilometer-notification-agent" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.151992 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="7da58715-94c1-487a-9f20-f008b4bea305" containerName="neutron-httpd" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.152010 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="sg-core" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.152025 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" containerName="proxy-httpd" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.159160 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.161848 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.162868 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.164950 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.172949 4936 scope.go:117] "RemoveContainer" containerID="d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.226849 4936 scope.go:117] "RemoveContainer" containerID="4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.249799 4936 scope.go:117] "RemoveContainer" containerID="001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522" Mar 20 16:20:31 crc kubenswrapper[4936]: E0320 16:20:31.255580 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522\": container with ID starting with 001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522 not found: ID does not exist" containerID="001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.255635 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522"} err="failed to get container status \"001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522\": rpc error: code = NotFound desc = could not find container \"001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522\": container with ID starting with 001202e36ec281237aadc77da9c801de8c4bb3e0333c754fbcc82f7708fac522 not found: ID does not exist" Mar 20 16:20:31 crc kubenswrapper[4936]: I0320 16:20:31.255667 4936 scope.go:117] "RemoveContainer" containerID="b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9" Mar 20 16:20:32 crc kubenswrapper[4936]: E0320 16:20:31.259924 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9\": container with ID starting with b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9 not found: ID does not exist" containerID="b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.259964 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9"} err="failed to get container status \"b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9\": rpc error: code = NotFound desc = could not find container \"b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9\": container with ID starting with b21c9018aa40b09937fe9c66c738e5e120c2705a6551ccf6c5fe7d74b6bfb0c9 not found: ID does not exist" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.259991 4936 scope.go:117] "RemoveContainer" containerID="d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c" Mar 20 16:20:32 crc kubenswrapper[4936]: E0320 16:20:31.260596 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c\": container with ID starting with d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c not found: ID does not exist" containerID="d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.260642 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c"} err="failed to get container status \"d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c\": rpc error: code = NotFound desc = could not find container \"d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c\": container with ID starting with d354f6e478e3c5acb59ce71bcadc867c1647cd1ed8df1cb5506e2aba6116897c not found: ID does not exist" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.260676 4936 scope.go:117] "RemoveContainer" containerID="4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35" Mar 20 16:20:32 crc kubenswrapper[4936]: E0320 16:20:31.262058 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35\": container with ID starting with 4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35 not found: ID does not exist" containerID="4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.262087 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35"} err="failed to get container status \"4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35\": rpc error: code = NotFound desc = could not find container \"4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35\": container with ID starting with 4d4769b06c843ffca6946bc3f90a7afe0a6db740186633680f82d0a5b9a8ce35 not found: ID does not exist" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.281428 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-log-httpd\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.281483 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-config-data\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.281520 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-run-httpd\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.281623 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.281656 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-scripts\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.281686 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.281724 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvh8l\" (UniqueName: \"kubernetes.io/projected/f25cfdda-1631-4199-9c75-88b95d557246-kube-api-access-fvh8l\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.383936 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-log-httpd\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.384005 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-config-data\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.384032 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-run-httpd\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.384086 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.384133 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-scripts\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.384158 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.384192 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvh8l\" (UniqueName: \"kubernetes.io/projected/f25cfdda-1631-4199-9c75-88b95d557246-kube-api-access-fvh8l\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.385118 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-log-httpd\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.385362 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-run-httpd\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.391402 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.391656 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-scripts\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.392250 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-config-data\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.393366 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.400966 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvh8l\" (UniqueName: \"kubernetes.io/projected/f25cfdda-1631-4199-9c75-88b95d557246-kube-api-access-fvh8l\") pod \"ceilometer-0\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.485985 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:31.874836 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c" path="/var/lib/kubelet/pods/d635a9cc-8b0e-45e4-b8ea-d81a34d74b5c/volumes" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:32.657972 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:32 crc kubenswrapper[4936]: W0320 16:20:32.670788 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf25cfdda_1631_4199_9c75_88b95d557246.slice/crio-2128aea10765d7bb05c9337de15ec41b4fea9ab12fe9b2fcf8ac948c5364b624 WatchSource:0}: Error finding container 2128aea10765d7bb05c9337de15ec41b4fea9ab12fe9b2fcf8ac948c5364b624: Status 404 returned error can't find the container with id 2128aea10765d7bb05c9337de15ec41b4fea9ab12fe9b2fcf8ac948c5364b624 Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:32.708788 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:32 crc kubenswrapper[4936]: I0320 16:20:32.739639 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-584c7d8894-w4d85" Mar 20 16:20:33 crc kubenswrapper[4936]: I0320 16:20:33.094682 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerStarted","Data":"2128aea10765d7bb05c9337de15ec41b4fea9ab12fe9b2fcf8ac948c5364b624"} Mar 20 16:20:34 crc kubenswrapper[4936]: I0320 16:20:34.110401 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerStarted","Data":"e380c2a25b483a3ea7d3ce2517d4974e2176ce51b6816b811734531f7cb8e348"} Mar 20 16:20:34 crc kubenswrapper[4936]: I0320 16:20:34.573918 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 20 16:20:37 crc kubenswrapper[4936]: I0320 16:20:37.343011 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:37 crc kubenswrapper[4936]: I0320 16:20:37.346423 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-74b5cc8579-wqhwv" Mar 20 16:20:38 crc kubenswrapper[4936]: I0320 16:20:38.316338 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:38 crc kubenswrapper[4936]: I0320 16:20:38.882234 4936 scope.go:117] "RemoveContainer" containerID="a37b6de81e8afb023b8879d76ef9385e8c34b13ed4f56cd1c86214ce40d0c5f3" Mar 20 16:20:39 crc kubenswrapper[4936]: I0320 16:20:39.850612 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79458c9dc6-vlzgk" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 20 16:20:39 crc kubenswrapper[4936]: I0320 16:20:39.850762 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:20:41 crc kubenswrapper[4936]: I0320 16:20:41.180866 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"961ea539-f955-46fd-8afa-8ec5b42133eb","Type":"ContainerStarted","Data":"11872c689031018f295292da405d7e082d515694493f1f7d15c66c9cc849d608"} Mar 20 16:20:41 crc kubenswrapper[4936]: I0320 16:20:41.183222 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerStarted","Data":"144c0c152b6a354bc6ee7f170dc31beb5638a58efc3a2962883baa9a893d7138"} Mar 20 16:20:41 crc kubenswrapper[4936]: I0320 16:20:41.201373 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.846194987 podStartE2EDuration="14.201352705s" podCreationTimestamp="2026-03-20 16:20:27 +0000 UTC" firstStartedPulling="2026-03-20 16:20:28.490605232 +0000 UTC m=+1179.436973047" lastFinishedPulling="2026-03-20 16:20:39.84576295 +0000 UTC m=+1190.792130765" observedRunningTime="2026-03-20 16:20:41.194124055 +0000 UTC m=+1192.140491870" watchObservedRunningTime="2026-03-20 16:20:41.201352705 +0000 UTC m=+1192.147720530" Mar 20 16:20:42 crc kubenswrapper[4936]: I0320 16:20:42.193822 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerStarted","Data":"f3becdc5ca9c72f0fa90c8bf73d7d35fc3bd1f91d921e17f59d337acdc9b2aee"} Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.550231 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-ktpms"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.556242 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.564885 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ktpms"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.606400 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wwjt\" (UniqueName: \"kubernetes.io/projected/04a90ca0-4946-4286-86b1-ba337895f3e8-kube-api-access-8wwjt\") pod \"nova-api-db-create-ktpms\" (UID: \"04a90ca0-4946-4286-86b1-ba337895f3e8\") " pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.606519 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a90ca0-4946-4286-86b1-ba337895f3e8-operator-scripts\") pod \"nova-api-db-create-ktpms\" (UID: \"04a90ca0-4946-4286-86b1-ba337895f3e8\") " pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.660812 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-f9qdx"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.663259 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.673326 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-f9qdx"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.707992 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-operator-scripts\") pod \"nova-cell0-db-create-f9qdx\" (UID: \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\") " pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.708058 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkl87\" (UniqueName: \"kubernetes.io/projected/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-kube-api-access-wkl87\") pod \"nova-cell0-db-create-f9qdx\" (UID: \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\") " pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.708121 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wwjt\" (UniqueName: \"kubernetes.io/projected/04a90ca0-4946-4286-86b1-ba337895f3e8-kube-api-access-8wwjt\") pod \"nova-api-db-create-ktpms\" (UID: \"04a90ca0-4946-4286-86b1-ba337895f3e8\") " pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.708169 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a90ca0-4946-4286-86b1-ba337895f3e8-operator-scripts\") pod \"nova-api-db-create-ktpms\" (UID: \"04a90ca0-4946-4286-86b1-ba337895f3e8\") " pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.709375 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a90ca0-4946-4286-86b1-ba337895f3e8-operator-scripts\") pod \"nova-api-db-create-ktpms\" (UID: \"04a90ca0-4946-4286-86b1-ba337895f3e8\") " pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.757374 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wwjt\" (UniqueName: \"kubernetes.io/projected/04a90ca0-4946-4286-86b1-ba337895f3e8-kube-api-access-8wwjt\") pod \"nova-api-db-create-ktpms\" (UID: \"04a90ca0-4946-4286-86b1-ba337895f3e8\") " pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.770663 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-06af-account-create-update-t8zxx"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.771906 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.777822 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.796353 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-06af-account-create-update-t8zxx"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.809460 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-operator-scripts\") pod \"nova-cell0-db-create-f9qdx\" (UID: \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\") " pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.809533 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9093776-e539-4e6e-ac2e-01da662cbe84-operator-scripts\") pod \"nova-api-06af-account-create-update-t8zxx\" (UID: \"d9093776-e539-4e6e-ac2e-01da662cbe84\") " pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.809585 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnxvt\" (UniqueName: \"kubernetes.io/projected/d9093776-e539-4e6e-ac2e-01da662cbe84-kube-api-access-bnxvt\") pod \"nova-api-06af-account-create-update-t8zxx\" (UID: \"d9093776-e539-4e6e-ac2e-01da662cbe84\") " pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.809606 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkl87\" (UniqueName: \"kubernetes.io/projected/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-kube-api-access-wkl87\") pod \"nova-cell0-db-create-f9qdx\" (UID: \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\") " pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.810920 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-operator-scripts\") pod \"nova-cell0-db-create-f9qdx\" (UID: \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\") " pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.846187 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkl87\" (UniqueName: \"kubernetes.io/projected/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-kube-api-access-wkl87\") pod \"nova-cell0-db-create-f9qdx\" (UID: \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\") " pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.865684 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-tv2ph"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.867122 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.877808 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tv2ph"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.911649 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9093776-e539-4e6e-ac2e-01da662cbe84-operator-scripts\") pod \"nova-api-06af-account-create-update-t8zxx\" (UID: \"d9093776-e539-4e6e-ac2e-01da662cbe84\") " pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.912402 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnxvt\" (UniqueName: \"kubernetes.io/projected/d9093776-e539-4e6e-ac2e-01da662cbe84-kube-api-access-bnxvt\") pod \"nova-api-06af-account-create-update-t8zxx\" (UID: \"d9093776-e539-4e6e-ac2e-01da662cbe84\") " pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.913366 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9093776-e539-4e6e-ac2e-01da662cbe84-operator-scripts\") pod \"nova-api-06af-account-create-update-t8zxx\" (UID: \"d9093776-e539-4e6e-ac2e-01da662cbe84\") " pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.946214 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnxvt\" (UniqueName: \"kubernetes.io/projected/d9093776-e539-4e6e-ac2e-01da662cbe84-kube-api-access-bnxvt\") pod \"nova-api-06af-account-create-update-t8zxx\" (UID: \"d9093776-e539-4e6e-ac2e-01da662cbe84\") " pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.967584 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-685b-account-create-update-j7zd9"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.968731 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.972846 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.977182 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-685b-account-create-update-j7zd9"] Mar 20 16:20:44 crc kubenswrapper[4936]: I0320 16:20:44.999135 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.007506 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.017380 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qs7m\" (UniqueName: \"kubernetes.io/projected/ec06ab44-bb2e-4add-8408-49d78727c812-kube-api-access-7qs7m\") pod \"nova-cell1-db-create-tv2ph\" (UID: \"ec06ab44-bb2e-4add-8408-49d78727c812\") " pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.017745 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-774r5\" (UniqueName: \"kubernetes.io/projected/f978ca59-3511-4a0f-9ae4-944e843f1c45-kube-api-access-774r5\") pod \"nova-cell0-685b-account-create-update-j7zd9\" (UID: \"f978ca59-3511-4a0f-9ae4-944e843f1c45\") " pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.017785 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec06ab44-bb2e-4add-8408-49d78727c812-operator-scripts\") pod \"nova-cell1-db-create-tv2ph\" (UID: \"ec06ab44-bb2e-4add-8408-49d78727c812\") " pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.017828 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f978ca59-3511-4a0f-9ae4-944e843f1c45-operator-scripts\") pod \"nova-cell0-685b-account-create-update-j7zd9\" (UID: \"f978ca59-3511-4a0f-9ae4-944e843f1c45\") " pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.122019 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec06ab44-bb2e-4add-8408-49d78727c812-operator-scripts\") pod \"nova-cell1-db-create-tv2ph\" (UID: \"ec06ab44-bb2e-4add-8408-49d78727c812\") " pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.122092 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f978ca59-3511-4a0f-9ae4-944e843f1c45-operator-scripts\") pod \"nova-cell0-685b-account-create-update-j7zd9\" (UID: \"f978ca59-3511-4a0f-9ae4-944e843f1c45\") " pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.122491 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qs7m\" (UniqueName: \"kubernetes.io/projected/ec06ab44-bb2e-4add-8408-49d78727c812-kube-api-access-7qs7m\") pod \"nova-cell1-db-create-tv2ph\" (UID: \"ec06ab44-bb2e-4add-8408-49d78727c812\") " pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.122574 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-774r5\" (UniqueName: \"kubernetes.io/projected/f978ca59-3511-4a0f-9ae4-944e843f1c45-kube-api-access-774r5\") pod \"nova-cell0-685b-account-create-update-j7zd9\" (UID: \"f978ca59-3511-4a0f-9ae4-944e843f1c45\") " pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.122663 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec06ab44-bb2e-4add-8408-49d78727c812-operator-scripts\") pod \"nova-cell1-db-create-tv2ph\" (UID: \"ec06ab44-bb2e-4add-8408-49d78727c812\") " pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.123209 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f978ca59-3511-4a0f-9ae4-944e843f1c45-operator-scripts\") pod \"nova-cell0-685b-account-create-update-j7zd9\" (UID: \"f978ca59-3511-4a0f-9ae4-944e843f1c45\") " pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.142815 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.147141 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qs7m\" (UniqueName: \"kubernetes.io/projected/ec06ab44-bb2e-4add-8408-49d78727c812-kube-api-access-7qs7m\") pod \"nova-cell1-db-create-tv2ph\" (UID: \"ec06ab44-bb2e-4add-8408-49d78727c812\") " pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.149630 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-774r5\" (UniqueName: \"kubernetes.io/projected/f978ca59-3511-4a0f-9ae4-944e843f1c45-kube-api-access-774r5\") pod \"nova-cell0-685b-account-create-update-j7zd9\" (UID: \"f978ca59-3511-4a0f-9ae4-944e843f1c45\") " pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.182738 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-416f-account-create-update-dk82t"] Mar 20 16:20:45 crc kubenswrapper[4936]: E0320 16:20:45.183203 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.183246 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon" Mar 20 16:20:45 crc kubenswrapper[4936]: E0320 16:20:45.183281 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon-log" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.183288 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon-log" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.183465 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.183476 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" containerName="horizon-log" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.184220 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.191003 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.196727 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-416f-account-create-update-dk82t"] Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.200299 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.209068 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.225068 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h85tg\" (UniqueName: \"kubernetes.io/projected/a4111fca-7b2f-456e-9591-e9f4b8f764a2-kube-api-access-h85tg\") pod \"nova-cell1-416f-account-create-update-dk82t\" (UID: \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\") " pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.225406 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4111fca-7b2f-456e-9591-e9f4b8f764a2-operator-scripts\") pod \"nova-cell1-416f-account-create-update-dk82t\" (UID: \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\") " pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.228807 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerStarted","Data":"72dd550d66bca8bb7250fb8f3d000fda5b20ab9825323e58a13e93b0fbb847ac"} Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.229007 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="ceilometer-central-agent" containerID="cri-o://e380c2a25b483a3ea7d3ce2517d4974e2176ce51b6816b811734531f7cb8e348" gracePeriod=30 Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.229306 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.229623 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="proxy-httpd" containerID="cri-o://72dd550d66bca8bb7250fb8f3d000fda5b20ab9825323e58a13e93b0fbb847ac" gracePeriod=30 Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.229680 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="sg-core" containerID="cri-o://f3becdc5ca9c72f0fa90c8bf73d7d35fc3bd1f91d921e17f59d337acdc9b2aee" gracePeriod=30 Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.229727 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="ceilometer-notification-agent" containerID="cri-o://144c0c152b6a354bc6ee7f170dc31beb5638a58efc3a2962883baa9a893d7138" gracePeriod=30 Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.233614 4936 generic.go:334] "Generic (PLEG): container finished" podID="02669555-c7e4-461a-86bd-aa30db944ab9" containerID="ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4" exitCode=137 Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.233669 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79458c9dc6-vlzgk" event={"ID":"02669555-c7e4-461a-86bd-aa30db944ab9","Type":"ContainerDied","Data":"ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4"} Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.233702 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79458c9dc6-vlzgk" event={"ID":"02669555-c7e4-461a-86bd-aa30db944ab9","Type":"ContainerDied","Data":"37f336d8f70e8fbe50557df41e8f7a1992229c8786b5c84e5558dc0a7a22f068"} Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.233721 4936 scope.go:117] "RemoveContainer" containerID="b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.233865 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79458c9dc6-vlzgk" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.262790 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.497718582 podStartE2EDuration="14.26276919s" podCreationTimestamp="2026-03-20 16:20:31 +0000 UTC" firstStartedPulling="2026-03-20 16:20:32.674465267 +0000 UTC m=+1183.620833082" lastFinishedPulling="2026-03-20 16:20:44.439515875 +0000 UTC m=+1195.385883690" observedRunningTime="2026-03-20 16:20:45.25480979 +0000 UTC m=+1196.201177605" watchObservedRunningTime="2026-03-20 16:20:45.26276919 +0000 UTC m=+1196.209136995" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.286385 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.327038 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-scripts\") pod \"02669555-c7e4-461a-86bd-aa30db944ab9\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.327158 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk2p7\" (UniqueName: \"kubernetes.io/projected/02669555-c7e4-461a-86bd-aa30db944ab9-kube-api-access-hk2p7\") pod \"02669555-c7e4-461a-86bd-aa30db944ab9\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.327188 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02669555-c7e4-461a-86bd-aa30db944ab9-logs\") pod \"02669555-c7e4-461a-86bd-aa30db944ab9\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.327237 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-tls-certs\") pod \"02669555-c7e4-461a-86bd-aa30db944ab9\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.327275 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-combined-ca-bundle\") pod \"02669555-c7e4-461a-86bd-aa30db944ab9\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.327330 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-config-data\") pod \"02669555-c7e4-461a-86bd-aa30db944ab9\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.327367 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-secret-key\") pod \"02669555-c7e4-461a-86bd-aa30db944ab9\" (UID: \"02669555-c7e4-461a-86bd-aa30db944ab9\") " Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.327693 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h85tg\" (UniqueName: \"kubernetes.io/projected/a4111fca-7b2f-456e-9591-e9f4b8f764a2-kube-api-access-h85tg\") pod \"nova-cell1-416f-account-create-update-dk82t\" (UID: \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\") " pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.327714 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4111fca-7b2f-456e-9591-e9f4b8f764a2-operator-scripts\") pod \"nova-cell1-416f-account-create-update-dk82t\" (UID: \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\") " pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.328566 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4111fca-7b2f-456e-9591-e9f4b8f764a2-operator-scripts\") pod \"nova-cell1-416f-account-create-update-dk82t\" (UID: \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\") " pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.337778 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02669555-c7e4-461a-86bd-aa30db944ab9-logs" (OuterVolumeSpecName: "logs") pod "02669555-c7e4-461a-86bd-aa30db944ab9" (UID: "02669555-c7e4-461a-86bd-aa30db944ab9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.357200 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "02669555-c7e4-461a-86bd-aa30db944ab9" (UID: "02669555-c7e4-461a-86bd-aa30db944ab9"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.365392 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02669555-c7e4-461a-86bd-aa30db944ab9-kube-api-access-hk2p7" (OuterVolumeSpecName: "kube-api-access-hk2p7") pod "02669555-c7e4-461a-86bd-aa30db944ab9" (UID: "02669555-c7e4-461a-86bd-aa30db944ab9"). InnerVolumeSpecName "kube-api-access-hk2p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.376167 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h85tg\" (UniqueName: \"kubernetes.io/projected/a4111fca-7b2f-456e-9591-e9f4b8f764a2-kube-api-access-h85tg\") pod \"nova-cell1-416f-account-create-update-dk82t\" (UID: \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\") " pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.387980 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-scripts" (OuterVolumeSpecName: "scripts") pod "02669555-c7e4-461a-86bd-aa30db944ab9" (UID: "02669555-c7e4-461a-86bd-aa30db944ab9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.414905 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02669555-c7e4-461a-86bd-aa30db944ab9" (UID: "02669555-c7e4-461a-86bd-aa30db944ab9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.416615 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-config-data" (OuterVolumeSpecName: "config-data") pod "02669555-c7e4-461a-86bd-aa30db944ab9" (UID: "02669555-c7e4-461a-86bd-aa30db944ab9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.426639 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "02669555-c7e4-461a-86bd-aa30db944ab9" (UID: "02669555-c7e4-461a-86bd-aa30db944ab9"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.429971 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.430011 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk2p7\" (UniqueName: \"kubernetes.io/projected/02669555-c7e4-461a-86bd-aa30db944ab9-kube-api-access-hk2p7\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.430022 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02669555-c7e4-461a-86bd-aa30db944ab9-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.430030 4936 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.430039 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.430047 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/02669555-c7e4-461a-86bd-aa30db944ab9-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.430056 4936 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/02669555-c7e4-461a-86bd-aa30db944ab9-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.489735 4936 scope.go:117] "RemoveContainer" containerID="ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.594871 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.612002 4936 scope.go:117] "RemoveContainer" containerID="b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9" Mar 20 16:20:45 crc kubenswrapper[4936]: E0320 16:20:45.615463 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9\": container with ID starting with b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9 not found: ID does not exist" containerID="b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.615510 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9"} err="failed to get container status \"b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9\": rpc error: code = NotFound desc = could not find container \"b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9\": container with ID starting with b1d635f50b28af67b9718db6c6498ec48169954e7d841101c3afa8ab8443ded9 not found: ID does not exist" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.615560 4936 scope.go:117] "RemoveContainer" containerID="ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.617507 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79458c9dc6-vlzgk"] Mar 20 16:20:45 crc kubenswrapper[4936]: E0320 16:20:45.618932 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4\": container with ID starting with ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4 not found: ID does not exist" containerID="ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.618993 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4"} err="failed to get container status \"ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4\": rpc error: code = NotFound desc = could not find container \"ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4\": container with ID starting with ec29884903563baf40770d8f1302f22dad1f7db138e0ba9e8657802e5d1c77d4 not found: ID does not exist" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.628893 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79458c9dc6-vlzgk"] Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.689628 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ktpms"] Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.825490 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-f9qdx"] Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.875450 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02669555-c7e4-461a-86bd-aa30db944ab9" path="/var/lib/kubelet/pods/02669555-c7e4-461a-86bd-aa30db944ab9/volumes" Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.889742 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-06af-account-create-update-t8zxx"] Mar 20 16:20:45 crc kubenswrapper[4936]: W0320 16:20:45.901200 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9093776_e539_4e6e_ac2e_01da662cbe84.slice/crio-102508c16713891a68ea335d9e654ff8958ba424bb1d362b718047d58d0f9e41 WatchSource:0}: Error finding container 102508c16713891a68ea335d9e654ff8958ba424bb1d362b718047d58d0f9e41: Status 404 returned error can't find the container with id 102508c16713891a68ea335d9e654ff8958ba424bb1d362b718047d58d0f9e41 Mar 20 16:20:45 crc kubenswrapper[4936]: I0320 16:20:45.927536 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tv2ph"] Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.124114 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-685b-account-create-update-j7zd9"] Mar 20 16:20:46 crc kubenswrapper[4936]: W0320 16:20:46.154954 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf978ca59_3511_4a0f_9ae4_944e843f1c45.slice/crio-2354cd66b3980015062b3e2aad3c20f0a972644d0d415d18270bfccd2f9e4a59 WatchSource:0}: Error finding container 2354cd66b3980015062b3e2aad3c20f0a972644d0d415d18270bfccd2f9e4a59: Status 404 returned error can't find the container with id 2354cd66b3980015062b3e2aad3c20f0a972644d0d415d18270bfccd2f9e4a59 Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.256121 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ktpms" event={"ID":"04a90ca0-4946-4286-86b1-ba337895f3e8","Type":"ContainerStarted","Data":"4ec4af133a8429106464a52c50ede79627d59cde86c3be8f1687b7ce31757ee7"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.256173 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ktpms" event={"ID":"04a90ca0-4946-4286-86b1-ba337895f3e8","Type":"ContainerStarted","Data":"081c3fd88b010b50d91bd27ff2d4c9597de2c5ddf9fcc2901c4b75291bb84cf1"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.268496 4936 generic.go:334] "Generic (PLEG): container finished" podID="f25cfdda-1631-4199-9c75-88b95d557246" containerID="72dd550d66bca8bb7250fb8f3d000fda5b20ab9825323e58a13e93b0fbb847ac" exitCode=0 Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.268535 4936 generic.go:334] "Generic (PLEG): container finished" podID="f25cfdda-1631-4199-9c75-88b95d557246" containerID="f3becdc5ca9c72f0fa90c8bf73d7d35fc3bd1f91d921e17f59d337acdc9b2aee" exitCode=2 Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.268546 4936 generic.go:334] "Generic (PLEG): container finished" podID="f25cfdda-1631-4199-9c75-88b95d557246" containerID="144c0c152b6a354bc6ee7f170dc31beb5638a58efc3a2962883baa9a893d7138" exitCode=0 Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.268566 4936 generic.go:334] "Generic (PLEG): container finished" podID="f25cfdda-1631-4199-9c75-88b95d557246" containerID="e380c2a25b483a3ea7d3ce2517d4974e2176ce51b6816b811734531f7cb8e348" exitCode=0 Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.268638 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerDied","Data":"72dd550d66bca8bb7250fb8f3d000fda5b20ab9825323e58a13e93b0fbb847ac"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.268671 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerDied","Data":"f3becdc5ca9c72f0fa90c8bf73d7d35fc3bd1f91d921e17f59d337acdc9b2aee"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.268682 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerDied","Data":"144c0c152b6a354bc6ee7f170dc31beb5638a58efc3a2962883baa9a893d7138"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.268691 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerDied","Data":"e380c2a25b483a3ea7d3ce2517d4974e2176ce51b6816b811734531f7cb8e348"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.288469 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-685b-account-create-update-j7zd9" event={"ID":"f978ca59-3511-4a0f-9ae4-944e843f1c45","Type":"ContainerStarted","Data":"2354cd66b3980015062b3e2aad3c20f0a972644d0d415d18270bfccd2f9e4a59"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.291190 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f9qdx" event={"ID":"ee8f6e3e-3bed-4aca-b7b4-98611a317a19","Type":"ContainerStarted","Data":"dd70f68a07ba25ef2d0a249aa2dd0e221a36c633e047a1871995e849f4664344"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.291221 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f9qdx" event={"ID":"ee8f6e3e-3bed-4aca-b7b4-98611a317a19","Type":"ContainerStarted","Data":"d6c56ac316cb2db2916ba744c32ec615c0cbb7c80630e62c7582ba81873484d7"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.294741 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-06af-account-create-update-t8zxx" event={"ID":"d9093776-e539-4e6e-ac2e-01da662cbe84","Type":"ContainerStarted","Data":"6590eaa90b73ad5b235781e69a2f1813aa5b5a16bbd87bb0d2552ced2ad89123"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.294795 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-06af-account-create-update-t8zxx" event={"ID":"d9093776-e539-4e6e-ac2e-01da662cbe84","Type":"ContainerStarted","Data":"102508c16713891a68ea335d9e654ff8958ba424bb1d362b718047d58d0f9e41"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.309709 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tv2ph" event={"ID":"ec06ab44-bb2e-4add-8408-49d78727c812","Type":"ContainerStarted","Data":"820e168264596bbee554652ab395adf1401820fca494d072f9a3ff01ec0d1db8"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.310095 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tv2ph" event={"ID":"ec06ab44-bb2e-4add-8408-49d78727c812","Type":"ContainerStarted","Data":"0c372330ed57e70b9f1a2fb404cb82342fc3359716909e82f0607364a929b4d8"} Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.312901 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.317895 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-f9qdx" podStartSLOduration=2.317881464 podStartE2EDuration="2.317881464s" podCreationTimestamp="2026-03-20 16:20:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:46.309933804 +0000 UTC m=+1197.256301619" watchObservedRunningTime="2026-03-20 16:20:46.317881464 +0000 UTC m=+1197.264249279" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.318534 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-416f-account-create-update-dk82t"] Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.329011 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-06af-account-create-update-t8zxx" podStartSLOduration=2.328993101 podStartE2EDuration="2.328993101s" podCreationTimestamp="2026-03-20 16:20:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:46.323262533 +0000 UTC m=+1197.269630348" watchObservedRunningTime="2026-03-20 16:20:46.328993101 +0000 UTC m=+1197.275360916" Mar 20 16:20:46 crc kubenswrapper[4936]: W0320 16:20:46.337831 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4111fca_7b2f_456e_9591_e9f4b8f764a2.slice/crio-9b581fcf9fcfb5909fe67a5c6ee98d8c54267c04e8ceb9a1dce32f2900360eba WatchSource:0}: Error finding container 9b581fcf9fcfb5909fe67a5c6ee98d8c54267c04e8ceb9a1dce32f2900360eba: Status 404 returned error can't find the container with id 9b581fcf9fcfb5909fe67a5c6ee98d8c54267c04e8ceb9a1dce32f2900360eba Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.346027 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-tv2ph" podStartSLOduration=2.3460100219999998 podStartE2EDuration="2.346010022s" podCreationTimestamp="2026-03-20 16:20:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:46.343203315 +0000 UTC m=+1197.289571140" watchObservedRunningTime="2026-03-20 16:20:46.346010022 +0000 UTC m=+1197.292377837" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.450366 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-log-httpd\") pod \"f25cfdda-1631-4199-9c75-88b95d557246\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.450480 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-config-data\") pod \"f25cfdda-1631-4199-9c75-88b95d557246\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.450559 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-combined-ca-bundle\") pod \"f25cfdda-1631-4199-9c75-88b95d557246\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.450605 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvh8l\" (UniqueName: \"kubernetes.io/projected/f25cfdda-1631-4199-9c75-88b95d557246-kube-api-access-fvh8l\") pod \"f25cfdda-1631-4199-9c75-88b95d557246\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.450643 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-run-httpd\") pod \"f25cfdda-1631-4199-9c75-88b95d557246\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.450675 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-sg-core-conf-yaml\") pod \"f25cfdda-1631-4199-9c75-88b95d557246\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.450768 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-scripts\") pod \"f25cfdda-1631-4199-9c75-88b95d557246\" (UID: \"f25cfdda-1631-4199-9c75-88b95d557246\") " Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.452242 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f25cfdda-1631-4199-9c75-88b95d557246" (UID: "f25cfdda-1631-4199-9c75-88b95d557246"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.452757 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f25cfdda-1631-4199-9c75-88b95d557246" (UID: "f25cfdda-1631-4199-9c75-88b95d557246"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.457523 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f25cfdda-1631-4199-9c75-88b95d557246-kube-api-access-fvh8l" (OuterVolumeSpecName: "kube-api-access-fvh8l") pod "f25cfdda-1631-4199-9c75-88b95d557246" (UID: "f25cfdda-1631-4199-9c75-88b95d557246"). InnerVolumeSpecName "kube-api-access-fvh8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.457768 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-scripts" (OuterVolumeSpecName: "scripts") pod "f25cfdda-1631-4199-9c75-88b95d557246" (UID: "f25cfdda-1631-4199-9c75-88b95d557246"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.515845 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f25cfdda-1631-4199-9c75-88b95d557246" (UID: "f25cfdda-1631-4199-9c75-88b95d557246"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.553837 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.553872 4936 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.553886 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvh8l\" (UniqueName: \"kubernetes.io/projected/f25cfdda-1631-4199-9c75-88b95d557246-kube-api-access-fvh8l\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.553901 4936 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25cfdda-1631-4199-9c75-88b95d557246-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.553911 4936 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.568378 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-config-data" (OuterVolumeSpecName: "config-data") pod "f25cfdda-1631-4199-9c75-88b95d557246" (UID: "f25cfdda-1631-4199-9c75-88b95d557246"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.569427 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f25cfdda-1631-4199-9c75-88b95d557246" (UID: "f25cfdda-1631-4199-9c75-88b95d557246"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.655178 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:46 crc kubenswrapper[4936]: I0320 16:20:46.655207 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25cfdda-1631-4199-9c75-88b95d557246-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.320835 4936 generic.go:334] "Generic (PLEG): container finished" podID="04a90ca0-4946-4286-86b1-ba337895f3e8" containerID="4ec4af133a8429106464a52c50ede79627d59cde86c3be8f1687b7ce31757ee7" exitCode=0 Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.320904 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ktpms" event={"ID":"04a90ca0-4946-4286-86b1-ba337895f3e8","Type":"ContainerDied","Data":"4ec4af133a8429106464a52c50ede79627d59cde86c3be8f1687b7ce31757ee7"} Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.325465 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25cfdda-1631-4199-9c75-88b95d557246","Type":"ContainerDied","Data":"2128aea10765d7bb05c9337de15ec41b4fea9ab12fe9b2fcf8ac948c5364b624"} Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.325507 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.325518 4936 scope.go:117] "RemoveContainer" containerID="72dd550d66bca8bb7250fb8f3d000fda5b20ab9825323e58a13e93b0fbb847ac" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.328133 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-685b-account-create-update-j7zd9" event={"ID":"f978ca59-3511-4a0f-9ae4-944e843f1c45","Type":"ContainerStarted","Data":"a1fda63ccf01efaf1986d551a0706a3f97e0f7ea62addcfe032a26f1d4882479"} Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.331496 4936 generic.go:334] "Generic (PLEG): container finished" podID="ee8f6e3e-3bed-4aca-b7b4-98611a317a19" containerID="dd70f68a07ba25ef2d0a249aa2dd0e221a36c633e047a1871995e849f4664344" exitCode=0 Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.331565 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f9qdx" event={"ID":"ee8f6e3e-3bed-4aca-b7b4-98611a317a19","Type":"ContainerDied","Data":"dd70f68a07ba25ef2d0a249aa2dd0e221a36c633e047a1871995e849f4664344"} Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.332820 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-416f-account-create-update-dk82t" event={"ID":"a4111fca-7b2f-456e-9591-e9f4b8f764a2","Type":"ContainerStarted","Data":"4eb396796326dbce4cf8a304e85aeb386499f7e4c7e8e4cdde3eab2678927f96"} Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.332848 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-416f-account-create-update-dk82t" event={"ID":"a4111fca-7b2f-456e-9591-e9f4b8f764a2","Type":"ContainerStarted","Data":"9b581fcf9fcfb5909fe67a5c6ee98d8c54267c04e8ceb9a1dce32f2900360eba"} Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.334673 4936 generic.go:334] "Generic (PLEG): container finished" podID="d9093776-e539-4e6e-ac2e-01da662cbe84" containerID="6590eaa90b73ad5b235781e69a2f1813aa5b5a16bbd87bb0d2552ced2ad89123" exitCode=0 Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.334740 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-06af-account-create-update-t8zxx" event={"ID":"d9093776-e539-4e6e-ac2e-01da662cbe84","Type":"ContainerDied","Data":"6590eaa90b73ad5b235781e69a2f1813aa5b5a16bbd87bb0d2552ced2ad89123"} Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.336448 4936 generic.go:334] "Generic (PLEG): container finished" podID="ec06ab44-bb2e-4add-8408-49d78727c812" containerID="820e168264596bbee554652ab395adf1401820fca494d072f9a3ff01ec0d1db8" exitCode=0 Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.336475 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tv2ph" event={"ID":"ec06ab44-bb2e-4add-8408-49d78727c812","Type":"ContainerDied","Data":"820e168264596bbee554652ab395adf1401820fca494d072f9a3ff01ec0d1db8"} Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.380055 4936 scope.go:117] "RemoveContainer" containerID="f3becdc5ca9c72f0fa90c8bf73d7d35fc3bd1f91d921e17f59d337acdc9b2aee" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.400115 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.409180 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.422826 4936 scope.go:117] "RemoveContainer" containerID="144c0c152b6a354bc6ee7f170dc31beb5638a58efc3a2962883baa9a893d7138" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.453683 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:47 crc kubenswrapper[4936]: E0320 16:20:47.454206 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="ceilometer-central-agent" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.454228 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="ceilometer-central-agent" Mar 20 16:20:47 crc kubenswrapper[4936]: E0320 16:20:47.454257 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="proxy-httpd" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.454264 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="proxy-httpd" Mar 20 16:20:47 crc kubenswrapper[4936]: E0320 16:20:47.454281 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="ceilometer-notification-agent" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.454287 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="ceilometer-notification-agent" Mar 20 16:20:47 crc kubenswrapper[4936]: E0320 16:20:47.454296 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="sg-core" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.454302 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="sg-core" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.454455 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="sg-core" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.454472 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="proxy-httpd" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.454495 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="ceilometer-notification-agent" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.454509 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f25cfdda-1631-4199-9c75-88b95d557246" containerName="ceilometer-central-agent" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.456477 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.460685 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.460835 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.482537 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-config-data\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.482686 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl8vz\" (UniqueName: \"kubernetes.io/projected/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-kube-api-access-cl8vz\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.482714 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-log-httpd\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.482787 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-run-httpd\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.482817 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.482846 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.482868 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-scripts\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.493946 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.497107 4936 scope.go:117] "RemoveContainer" containerID="e380c2a25b483a3ea7d3ce2517d4974e2176ce51b6816b811734531f7cb8e348" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.514519 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-416f-account-create-update-dk82t" podStartSLOduration=2.514441845 podStartE2EDuration="2.514441845s" podCreationTimestamp="2026-03-20 16:20:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:47.463450073 +0000 UTC m=+1198.409817888" watchObservedRunningTime="2026-03-20 16:20:47.514441845 +0000 UTC m=+1198.460809680" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.520966 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-685b-account-create-update-j7zd9" podStartSLOduration=3.520946705 podStartE2EDuration="3.520946705s" podCreationTimestamp="2026-03-20 16:20:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:20:47.475902408 +0000 UTC m=+1198.422270223" watchObservedRunningTime="2026-03-20 16:20:47.520946705 +0000 UTC m=+1198.467314520" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.584663 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-config-data\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.584746 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl8vz\" (UniqueName: \"kubernetes.io/projected/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-kube-api-access-cl8vz\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.584767 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-log-httpd\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.584819 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-run-httpd\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.584839 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.584860 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.584891 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-scripts\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.586843 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-log-httpd\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.590040 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-scripts\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.590285 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-run-httpd\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.592355 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.605355 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.605776 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-config-data\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.612414 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl8vz\" (UniqueName: \"kubernetes.io/projected/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-kube-api-access-cl8vz\") pod \"ceilometer-0\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.795534 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.867037 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f25cfdda-1631-4199-9c75-88b95d557246" path="/var/lib/kubelet/pods/f25cfdda-1631-4199-9c75-88b95d557246/volumes" Mar 20 16:20:47 crc kubenswrapper[4936]: I0320 16:20:47.947490 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.093462 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wwjt\" (UniqueName: \"kubernetes.io/projected/04a90ca0-4946-4286-86b1-ba337895f3e8-kube-api-access-8wwjt\") pod \"04a90ca0-4946-4286-86b1-ba337895f3e8\" (UID: \"04a90ca0-4946-4286-86b1-ba337895f3e8\") " Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.093896 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a90ca0-4946-4286-86b1-ba337895f3e8-operator-scripts\") pod \"04a90ca0-4946-4286-86b1-ba337895f3e8\" (UID: \"04a90ca0-4946-4286-86b1-ba337895f3e8\") " Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.095268 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04a90ca0-4946-4286-86b1-ba337895f3e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "04a90ca0-4946-4286-86b1-ba337895f3e8" (UID: "04a90ca0-4946-4286-86b1-ba337895f3e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.099948 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04a90ca0-4946-4286-86b1-ba337895f3e8-kube-api-access-8wwjt" (OuterVolumeSpecName: "kube-api-access-8wwjt") pod "04a90ca0-4946-4286-86b1-ba337895f3e8" (UID: "04a90ca0-4946-4286-86b1-ba337895f3e8"). InnerVolumeSpecName "kube-api-access-8wwjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.196079 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wwjt\" (UniqueName: \"kubernetes.io/projected/04a90ca0-4946-4286-86b1-ba337895f3e8-kube-api-access-8wwjt\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.196114 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a90ca0-4946-4286-86b1-ba337895f3e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.300353 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.348636 4936 generic.go:334] "Generic (PLEG): container finished" podID="f978ca59-3511-4a0f-9ae4-944e843f1c45" containerID="a1fda63ccf01efaf1986d551a0706a3f97e0f7ea62addcfe032a26f1d4882479" exitCode=0 Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.349270 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-685b-account-create-update-j7zd9" event={"ID":"f978ca59-3511-4a0f-9ae4-944e843f1c45","Type":"ContainerDied","Data":"a1fda63ccf01efaf1986d551a0706a3f97e0f7ea62addcfe032a26f1d4882479"} Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.352274 4936 generic.go:334] "Generic (PLEG): container finished" podID="a4111fca-7b2f-456e-9591-e9f4b8f764a2" containerID="4eb396796326dbce4cf8a304e85aeb386499f7e4c7e8e4cdde3eab2678927f96" exitCode=0 Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.352455 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-416f-account-create-update-dk82t" event={"ID":"a4111fca-7b2f-456e-9591-e9f4b8f764a2","Type":"ContainerDied","Data":"4eb396796326dbce4cf8a304e85aeb386499f7e4c7e8e4cdde3eab2678927f96"} Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.354633 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerStarted","Data":"62f3a43acf9469b6d420cfc6d89003e747fe55f03fe06cee66d239a222a8e956"} Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.355797 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ktpms" event={"ID":"04a90ca0-4946-4286-86b1-ba337895f3e8","Type":"ContainerDied","Data":"081c3fd88b010b50d91bd27ff2d4c9597de2c5ddf9fcc2901c4b75291bb84cf1"} Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.355818 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="081c3fd88b010b50d91bd27ff2d4c9597de2c5ddf9fcc2901c4b75291bb84cf1" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.355866 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ktpms" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.883849 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.899578 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.910211 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.911023 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkl87\" (UniqueName: \"kubernetes.io/projected/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-kube-api-access-wkl87\") pod \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\" (UID: \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\") " Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.911190 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec06ab44-bb2e-4add-8408-49d78727c812-operator-scripts\") pod \"ec06ab44-bb2e-4add-8408-49d78727c812\" (UID: \"ec06ab44-bb2e-4add-8408-49d78727c812\") " Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.911276 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-operator-scripts\") pod \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\" (UID: \"ee8f6e3e-3bed-4aca-b7b4-98611a317a19\") " Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.911377 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qs7m\" (UniqueName: \"kubernetes.io/projected/ec06ab44-bb2e-4add-8408-49d78727c812-kube-api-access-7qs7m\") pod \"ec06ab44-bb2e-4add-8408-49d78727c812\" (UID: \"ec06ab44-bb2e-4add-8408-49d78727c812\") " Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.915971 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee8f6e3e-3bed-4aca-b7b4-98611a317a19" (UID: "ee8f6e3e-3bed-4aca-b7b4-98611a317a19"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.916282 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec06ab44-bb2e-4add-8408-49d78727c812-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec06ab44-bb2e-4add-8408-49d78727c812" (UID: "ec06ab44-bb2e-4add-8408-49d78727c812"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.950664 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec06ab44-bb2e-4add-8408-49d78727c812-kube-api-access-7qs7m" (OuterVolumeSpecName: "kube-api-access-7qs7m") pod "ec06ab44-bb2e-4add-8408-49d78727c812" (UID: "ec06ab44-bb2e-4add-8408-49d78727c812"). InnerVolumeSpecName "kube-api-access-7qs7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:48 crc kubenswrapper[4936]: I0320 16:20:48.950785 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-kube-api-access-wkl87" (OuterVolumeSpecName: "kube-api-access-wkl87") pod "ee8f6e3e-3bed-4aca-b7b4-98611a317a19" (UID: "ee8f6e3e-3bed-4aca-b7b4-98611a317a19"). InnerVolumeSpecName "kube-api-access-wkl87". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.014326 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9093776-e539-4e6e-ac2e-01da662cbe84-operator-scripts\") pod \"d9093776-e539-4e6e-ac2e-01da662cbe84\" (UID: \"d9093776-e539-4e6e-ac2e-01da662cbe84\") " Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.014634 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnxvt\" (UniqueName: \"kubernetes.io/projected/d9093776-e539-4e6e-ac2e-01da662cbe84-kube-api-access-bnxvt\") pod \"d9093776-e539-4e6e-ac2e-01da662cbe84\" (UID: \"d9093776-e539-4e6e-ac2e-01da662cbe84\") " Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.015301 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkl87\" (UniqueName: \"kubernetes.io/projected/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-kube-api-access-wkl87\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.015333 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec06ab44-bb2e-4add-8408-49d78727c812-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.015348 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8f6e3e-3bed-4aca-b7b4-98611a317a19-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.015360 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qs7m\" (UniqueName: \"kubernetes.io/projected/ec06ab44-bb2e-4add-8408-49d78727c812-kube-api-access-7qs7m\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.015737 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9093776-e539-4e6e-ac2e-01da662cbe84-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d9093776-e539-4e6e-ac2e-01da662cbe84" (UID: "d9093776-e539-4e6e-ac2e-01da662cbe84"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.019273 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9093776-e539-4e6e-ac2e-01da662cbe84-kube-api-access-bnxvt" (OuterVolumeSpecName: "kube-api-access-bnxvt") pod "d9093776-e539-4e6e-ac2e-01da662cbe84" (UID: "d9093776-e539-4e6e-ac2e-01da662cbe84"). InnerVolumeSpecName "kube-api-access-bnxvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.116478 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9093776-e539-4e6e-ac2e-01da662cbe84-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.116509 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnxvt\" (UniqueName: \"kubernetes.io/projected/d9093776-e539-4e6e-ac2e-01da662cbe84-kube-api-access-bnxvt\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.398129 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f9qdx" event={"ID":"ee8f6e3e-3bed-4aca-b7b4-98611a317a19","Type":"ContainerDied","Data":"d6c56ac316cb2db2916ba744c32ec615c0cbb7c80630e62c7582ba81873484d7"} Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.398387 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6c56ac316cb2db2916ba744c32ec615c0cbb7c80630e62c7582ba81873484d7" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.398461 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f9qdx" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.417165 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-06af-account-create-update-t8zxx" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.419415 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-06af-account-create-update-t8zxx" event={"ID":"d9093776-e539-4e6e-ac2e-01da662cbe84","Type":"ContainerDied","Data":"102508c16713891a68ea335d9e654ff8958ba424bb1d362b718047d58d0f9e41"} Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.419460 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="102508c16713891a68ea335d9e654ff8958ba424bb1d362b718047d58d0f9e41" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.446998 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tv2ph" event={"ID":"ec06ab44-bb2e-4add-8408-49d78727c812","Type":"ContainerDied","Data":"0c372330ed57e70b9f1a2fb404cb82342fc3359716909e82f0607364a929b4d8"} Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.447048 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c372330ed57e70b9f1a2fb404cb82342fc3359716909e82f0607364a929b4d8" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.447140 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tv2ph" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.958517 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:49 crc kubenswrapper[4936]: I0320 16:20:49.975562 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.117816 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.150128 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h85tg\" (UniqueName: \"kubernetes.io/projected/a4111fca-7b2f-456e-9591-e9f4b8f764a2-kube-api-access-h85tg\") pod \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\" (UID: \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\") " Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.150294 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4111fca-7b2f-456e-9591-e9f4b8f764a2-operator-scripts\") pod \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\" (UID: \"a4111fca-7b2f-456e-9591-e9f4b8f764a2\") " Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.150357 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f978ca59-3511-4a0f-9ae4-944e843f1c45-operator-scripts\") pod \"f978ca59-3511-4a0f-9ae4-944e843f1c45\" (UID: \"f978ca59-3511-4a0f-9ae4-944e843f1c45\") " Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.150429 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-774r5\" (UniqueName: \"kubernetes.io/projected/f978ca59-3511-4a0f-9ae4-944e843f1c45-kube-api-access-774r5\") pod \"f978ca59-3511-4a0f-9ae4-944e843f1c45\" (UID: \"f978ca59-3511-4a0f-9ae4-944e843f1c45\") " Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.152148 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4111fca-7b2f-456e-9591-e9f4b8f764a2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a4111fca-7b2f-456e-9591-e9f4b8f764a2" (UID: "a4111fca-7b2f-456e-9591-e9f4b8f764a2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.155482 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f978ca59-3511-4a0f-9ae4-944e843f1c45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f978ca59-3511-4a0f-9ae4-944e843f1c45" (UID: "f978ca59-3511-4a0f-9ae4-944e843f1c45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.170481 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4111fca-7b2f-456e-9591-e9f4b8f764a2-kube-api-access-h85tg" (OuterVolumeSpecName: "kube-api-access-h85tg") pod "a4111fca-7b2f-456e-9591-e9f4b8f764a2" (UID: "a4111fca-7b2f-456e-9591-e9f4b8f764a2"). InnerVolumeSpecName "kube-api-access-h85tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.171537 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f978ca59-3511-4a0f-9ae4-944e843f1c45-kube-api-access-774r5" (OuterVolumeSpecName: "kube-api-access-774r5") pod "f978ca59-3511-4a0f-9ae4-944e843f1c45" (UID: "f978ca59-3511-4a0f-9ae4-944e843f1c45"). InnerVolumeSpecName "kube-api-access-774r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.252643 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h85tg\" (UniqueName: \"kubernetes.io/projected/a4111fca-7b2f-456e-9591-e9f4b8f764a2-kube-api-access-h85tg\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.252688 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4111fca-7b2f-456e-9591-e9f4b8f764a2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.252701 4936 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f978ca59-3511-4a0f-9ae4-944e843f1c45-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.252713 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-774r5\" (UniqueName: \"kubernetes.io/projected/f978ca59-3511-4a0f-9ae4-944e843f1c45-kube-api-access-774r5\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.456496 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerStarted","Data":"67edf8e1cb1d22c35ab777acf82f0ed73c5d38db76eaee80ee44e593fe90e2a8"} Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.456607 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerStarted","Data":"f24d804fe426ef9774fe50735a24ff592c9edf9fd2fb97e9e6b77c6bbbe59e54"} Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.459948 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-685b-account-create-update-j7zd9" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.459946 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-685b-account-create-update-j7zd9" event={"ID":"f978ca59-3511-4a0f-9ae4-944e843f1c45","Type":"ContainerDied","Data":"2354cd66b3980015062b3e2aad3c20f0a972644d0d415d18270bfccd2f9e4a59"} Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.460082 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2354cd66b3980015062b3e2aad3c20f0a972644d0d415d18270bfccd2f9e4a59" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.467731 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-416f-account-create-update-dk82t" event={"ID":"a4111fca-7b2f-456e-9591-e9f4b8f764a2","Type":"ContainerDied","Data":"9b581fcf9fcfb5909fe67a5c6ee98d8c54267c04e8ceb9a1dce32f2900360eba"} Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.467779 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b581fcf9fcfb5909fe67a5c6ee98d8c54267c04e8ceb9a1dce32f2900360eba" Mar 20 16:20:50 crc kubenswrapper[4936]: I0320 16:20:50.467839 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-416f-account-create-update-dk82t" Mar 20 16:20:51 crc kubenswrapper[4936]: I0320 16:20:51.477102 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerStarted","Data":"805bbe8ba0efb48806e5c8dbf2968acd687d1f78ef4fd0f3c5f52b811a392124"} Mar 20 16:20:54 crc kubenswrapper[4936]: I0320 16:20:54.501204 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerStarted","Data":"2b5ba4f2f3a5f173aa66dbc1fcd2e3b9fce7a851b02f314f2ddf270371e3904c"} Mar 20 16:20:54 crc kubenswrapper[4936]: I0320 16:20:54.501946 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="ceilometer-central-agent" containerID="cri-o://f24d804fe426ef9774fe50735a24ff592c9edf9fd2fb97e9e6b77c6bbbe59e54" gracePeriod=30 Mar 20 16:20:54 crc kubenswrapper[4936]: I0320 16:20:54.502236 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 20 16:20:54 crc kubenswrapper[4936]: I0320 16:20:54.502500 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="proxy-httpd" containerID="cri-o://2b5ba4f2f3a5f173aa66dbc1fcd2e3b9fce7a851b02f314f2ddf270371e3904c" gracePeriod=30 Mar 20 16:20:54 crc kubenswrapper[4936]: I0320 16:20:54.502558 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="sg-core" containerID="cri-o://805bbe8ba0efb48806e5c8dbf2968acd687d1f78ef4fd0f3c5f52b811a392124" gracePeriod=30 Mar 20 16:20:54 crc kubenswrapper[4936]: I0320 16:20:54.502593 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="ceilometer-notification-agent" containerID="cri-o://67edf8e1cb1d22c35ab777acf82f0ed73c5d38db76eaee80ee44e593fe90e2a8" gracePeriod=30 Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.279017 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.965965394 podStartE2EDuration="8.278986934s" podCreationTimestamp="2026-03-20 16:20:47 +0000 UTC" firstStartedPulling="2026-03-20 16:20:48.305130508 +0000 UTC m=+1199.251498323" lastFinishedPulling="2026-03-20 16:20:53.618152048 +0000 UTC m=+1204.564519863" observedRunningTime="2026-03-20 16:20:54.531942059 +0000 UTC m=+1205.478309944" watchObservedRunningTime="2026-03-20 16:20:55.278986934 +0000 UTC m=+1206.225354749" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281065 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hspgq"] Mar 20 16:20:55 crc kubenswrapper[4936]: E0320 16:20:55.281533 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04a90ca0-4946-4286-86b1-ba337895f3e8" containerName="mariadb-database-create" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281571 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="04a90ca0-4946-4286-86b1-ba337895f3e8" containerName="mariadb-database-create" Mar 20 16:20:55 crc kubenswrapper[4936]: E0320 16:20:55.281592 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee8f6e3e-3bed-4aca-b7b4-98611a317a19" containerName="mariadb-database-create" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281600 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee8f6e3e-3bed-4aca-b7b4-98611a317a19" containerName="mariadb-database-create" Mar 20 16:20:55 crc kubenswrapper[4936]: E0320 16:20:55.281616 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec06ab44-bb2e-4add-8408-49d78727c812" containerName="mariadb-database-create" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281625 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec06ab44-bb2e-4add-8408-49d78727c812" containerName="mariadb-database-create" Mar 20 16:20:55 crc kubenswrapper[4936]: E0320 16:20:55.281637 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4111fca-7b2f-456e-9591-e9f4b8f764a2" containerName="mariadb-account-create-update" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281645 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4111fca-7b2f-456e-9591-e9f4b8f764a2" containerName="mariadb-account-create-update" Mar 20 16:20:55 crc kubenswrapper[4936]: E0320 16:20:55.281658 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f978ca59-3511-4a0f-9ae4-944e843f1c45" containerName="mariadb-account-create-update" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281665 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f978ca59-3511-4a0f-9ae4-944e843f1c45" containerName="mariadb-account-create-update" Mar 20 16:20:55 crc kubenswrapper[4936]: E0320 16:20:55.281684 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9093776-e539-4e6e-ac2e-01da662cbe84" containerName="mariadb-account-create-update" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281691 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9093776-e539-4e6e-ac2e-01da662cbe84" containerName="mariadb-account-create-update" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281921 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="04a90ca0-4946-4286-86b1-ba337895f3e8" containerName="mariadb-database-create" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281947 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4111fca-7b2f-456e-9591-e9f4b8f764a2" containerName="mariadb-account-create-update" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281964 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee8f6e3e-3bed-4aca-b7b4-98611a317a19" containerName="mariadb-database-create" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281976 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f978ca59-3511-4a0f-9ae4-944e843f1c45" containerName="mariadb-account-create-update" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281987 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9093776-e539-4e6e-ac2e-01da662cbe84" containerName="mariadb-account-create-update" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.281997 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec06ab44-bb2e-4add-8408-49d78727c812" containerName="mariadb-database-create" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.282805 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.284587 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.284900 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.285130 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-jmd6p" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.289655 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hspgq"] Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.445016 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2djg\" (UniqueName: \"kubernetes.io/projected/49bf6254-7f19-41f0-969b-fc8b338ece00-kube-api-access-q2djg\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.445085 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-scripts\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.445141 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-config-data\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.445207 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.513476 4936 generic.go:334] "Generic (PLEG): container finished" podID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerID="2b5ba4f2f3a5f173aa66dbc1fcd2e3b9fce7a851b02f314f2ddf270371e3904c" exitCode=0 Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.513512 4936 generic.go:334] "Generic (PLEG): container finished" podID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerID="805bbe8ba0efb48806e5c8dbf2968acd687d1f78ef4fd0f3c5f52b811a392124" exitCode=2 Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.513521 4936 generic.go:334] "Generic (PLEG): container finished" podID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerID="67edf8e1cb1d22c35ab777acf82f0ed73c5d38db76eaee80ee44e593fe90e2a8" exitCode=0 Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.513540 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerDied","Data":"2b5ba4f2f3a5f173aa66dbc1fcd2e3b9fce7a851b02f314f2ddf270371e3904c"} Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.513578 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerDied","Data":"805bbe8ba0efb48806e5c8dbf2968acd687d1f78ef4fd0f3c5f52b811a392124"} Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.513588 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerDied","Data":"67edf8e1cb1d22c35ab777acf82f0ed73c5d38db76eaee80ee44e593fe90e2a8"} Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.547251 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2djg\" (UniqueName: \"kubernetes.io/projected/49bf6254-7f19-41f0-969b-fc8b338ece00-kube-api-access-q2djg\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.547353 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-scripts\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.548294 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-config-data\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.548380 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.553436 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.565369 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2djg\" (UniqueName: \"kubernetes.io/projected/49bf6254-7f19-41f0-969b-fc8b338ece00-kube-api-access-q2djg\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.565858 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-scripts\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.566720 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-config-data\") pod \"nova-cell0-conductor-db-sync-hspgq\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:55 crc kubenswrapper[4936]: I0320 16:20:55.601462 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:20:56 crc kubenswrapper[4936]: I0320 16:20:56.055098 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hspgq"] Mar 20 16:20:56 crc kubenswrapper[4936]: W0320 16:20:56.076848 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49bf6254_7f19_41f0_969b_fc8b338ece00.slice/crio-e777ea75fe121c555066848cd2b59a9d28426b9cc6c251ba6c3ef4f656cbaf15 WatchSource:0}: Error finding container e777ea75fe121c555066848cd2b59a9d28426b9cc6c251ba6c3ef4f656cbaf15: Status 404 returned error can't find the container with id e777ea75fe121c555066848cd2b59a9d28426b9cc6c251ba6c3ef4f656cbaf15 Mar 20 16:20:56 crc kubenswrapper[4936]: I0320 16:20:56.528007 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hspgq" event={"ID":"49bf6254-7f19-41f0-969b-fc8b338ece00","Type":"ContainerStarted","Data":"e777ea75fe121c555066848cd2b59a9d28426b9cc6c251ba6c3ef4f656cbaf15"} Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.550916 4936 generic.go:334] "Generic (PLEG): container finished" podID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerID="f24d804fe426ef9774fe50735a24ff592c9edf9fd2fb97e9e6b77c6bbbe59e54" exitCode=0 Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.551442 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerDied","Data":"f24d804fe426ef9774fe50735a24ff592c9edf9fd2fb97e9e6b77c6bbbe59e54"} Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.680956 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.813845 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-scripts\") pod \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.814315 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-config-data\") pod \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.814477 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-sg-core-conf-yaml\") pod \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.814669 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-run-httpd\") pod \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.814903 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-combined-ca-bundle\") pod \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.815106 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl8vz\" (UniqueName: \"kubernetes.io/projected/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-kube-api-access-cl8vz\") pod \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.815321 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-log-httpd\") pod \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\" (UID: \"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3\") " Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.815666 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" (UID: "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.816174 4936 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.816332 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" (UID: "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.821784 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-kube-api-access-cl8vz" (OuterVolumeSpecName: "kube-api-access-cl8vz") pod "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" (UID: "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3"). InnerVolumeSpecName "kube-api-access-cl8vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.822721 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-scripts" (OuterVolumeSpecName: "scripts") pod "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" (UID: "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.847430 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" (UID: "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.898591 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" (UID: "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.920406 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.920444 4936 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.920454 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.920463 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl8vz\" (UniqueName: \"kubernetes.io/projected/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-kube-api-access-cl8vz\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.920473 4936 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:58 crc kubenswrapper[4936]: I0320 16:20:58.932764 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-config-data" (OuterVolumeSpecName: "config-data") pod "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" (UID: "1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.022912 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.566567 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3","Type":"ContainerDied","Data":"62f3a43acf9469b6d420cfc6d89003e747fe55f03fe06cee66d239a222a8e956"} Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.566880 4936 scope.go:117] "RemoveContainer" containerID="2b5ba4f2f3a5f173aa66dbc1fcd2e3b9fce7a851b02f314f2ddf270371e3904c" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.566740 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.590084 4936 scope.go:117] "RemoveContainer" containerID="805bbe8ba0efb48806e5c8dbf2968acd687d1f78ef4fd0f3c5f52b811a392124" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.612342 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.626518 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.643797 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:59 crc kubenswrapper[4936]: E0320 16:20:59.644218 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="ceilometer-notification-agent" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.644242 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="ceilometer-notification-agent" Mar 20 16:20:59 crc kubenswrapper[4936]: E0320 16:20:59.644252 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="proxy-httpd" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.644261 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="proxy-httpd" Mar 20 16:20:59 crc kubenswrapper[4936]: E0320 16:20:59.644288 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="sg-core" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.644296 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="sg-core" Mar 20 16:20:59 crc kubenswrapper[4936]: E0320 16:20:59.644311 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="ceilometer-central-agent" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.644319 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="ceilometer-central-agent" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.648438 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="ceilometer-notification-agent" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.648476 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="proxy-httpd" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.648488 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="sg-core" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.648513 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" containerName="ceilometer-central-agent" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.650123 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.654029 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.654293 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.659035 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.840274 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-config-data\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.840414 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-scripts\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.840607 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-run-httpd\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.840709 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44wl9\" (UniqueName: \"kubernetes.io/projected/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-kube-api-access-44wl9\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.840769 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.841035 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.841134 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-log-httpd\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.942866 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-run-httpd\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.942981 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44wl9\" (UniqueName: \"kubernetes.io/projected/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-kube-api-access-44wl9\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.943015 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.943151 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.943213 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-log-httpd\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.943250 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-config-data\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.943293 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-scripts\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.943759 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-run-httpd\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.944301 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-log-httpd\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.949011 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.949474 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-scripts\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.949717 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.953409 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3" path="/var/lib/kubelet/pods/1650d2f1-ac76-4749-bf40-b0d6ce1ccaf3/volumes" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.968718 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-config-data\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.976007 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44wl9\" (UniqueName: \"kubernetes.io/projected/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-kube-api-access-44wl9\") pod \"ceilometer-0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " pod="openstack/ceilometer-0" Mar 20 16:20:59 crc kubenswrapper[4936]: I0320 16:20:59.979920 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:21:00 crc kubenswrapper[4936]: I0320 16:21:00.389751 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:21:00 crc kubenswrapper[4936]: I0320 16:21:00.390051 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerName="glance-log" containerID="cri-o://9433d1ec221bb76adc3c454f903b2634b9e3a2292e74d1bbd6adb254492f958d" gracePeriod=30 Mar 20 16:21:00 crc kubenswrapper[4936]: I0320 16:21:00.390159 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerName="glance-httpd" containerID="cri-o://1ca8e48a473928080395ecdceb6e5c0280a91455aee8c60600269c60c57486d4" gracePeriod=30 Mar 20 16:21:00 crc kubenswrapper[4936]: I0320 16:21:00.585221 4936 generic.go:334] "Generic (PLEG): container finished" podID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerID="9433d1ec221bb76adc3c454f903b2634b9e3a2292e74d1bbd6adb254492f958d" exitCode=143 Mar 20 16:21:00 crc kubenswrapper[4936]: I0320 16:21:00.585294 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be14f196-9b19-49e3-bc07-b642c9ae7ff8","Type":"ContainerDied","Data":"9433d1ec221bb76adc3c454f903b2634b9e3a2292e74d1bbd6adb254492f958d"} Mar 20 16:21:01 crc kubenswrapper[4936]: I0320 16:21:01.117006 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:21:01 crc kubenswrapper[4936]: I0320 16:21:01.120040 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerName="glance-log" containerID="cri-o://5d07190ffa6702b331b01eafb1e8f1975d312c0eb6d60b7f88fce924ab3951af" gracePeriod=30 Mar 20 16:21:01 crc kubenswrapper[4936]: I0320 16:21:01.120124 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerName="glance-httpd" containerID="cri-o://b4d22dfeb13ddda6b999a4c43aab3870c668690858a2a94abfcfdc556416edd3" gracePeriod=30 Mar 20 16:21:01 crc kubenswrapper[4936]: I0320 16:21:01.539674 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:01 crc kubenswrapper[4936]: I0320 16:21:01.597828 4936 generic.go:334] "Generic (PLEG): container finished" podID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerID="5d07190ffa6702b331b01eafb1e8f1975d312c0eb6d60b7f88fce924ab3951af" exitCode=143 Mar 20 16:21:01 crc kubenswrapper[4936]: I0320 16:21:01.597874 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"601c14d9-f6a4-4f7c-b033-7a35d41594a2","Type":"ContainerDied","Data":"5d07190ffa6702b331b01eafb1e8f1975d312c0eb6d60b7f88fce924ab3951af"} Mar 20 16:21:04 crc kubenswrapper[4936]: I0320 16:21:04.240982 4936 scope.go:117] "RemoveContainer" containerID="67edf8e1cb1d22c35ab777acf82f0ed73c5d38db76eaee80ee44e593fe90e2a8" Mar 20 16:21:04 crc kubenswrapper[4936]: I0320 16:21:04.628959 4936 generic.go:334] "Generic (PLEG): container finished" podID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerID="1ca8e48a473928080395ecdceb6e5c0280a91455aee8c60600269c60c57486d4" exitCode=0 Mar 20 16:21:04 crc kubenswrapper[4936]: I0320 16:21:04.629028 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be14f196-9b19-49e3-bc07-b642c9ae7ff8","Type":"ContainerDied","Data":"1ca8e48a473928080395ecdceb6e5c0280a91455aee8c60600269c60c57486d4"} Mar 20 16:21:04 crc kubenswrapper[4936]: I0320 16:21:04.631417 4936 generic.go:334] "Generic (PLEG): container finished" podID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerID="b4d22dfeb13ddda6b999a4c43aab3870c668690858a2a94abfcfdc556416edd3" exitCode=0 Mar 20 16:21:04 crc kubenswrapper[4936]: I0320 16:21:04.631484 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"601c14d9-f6a4-4f7c-b033-7a35d41594a2","Type":"ContainerDied","Data":"b4d22dfeb13ddda6b999a4c43aab3870c668690858a2a94abfcfdc556416edd3"} Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.765631 4936 scope.go:117] "RemoveContainer" containerID="f24d804fe426ef9774fe50735a24ff592c9edf9fd2fb97e9e6b77c6bbbe59e54" Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.937198 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.962948 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-config-data\") pod \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.963047 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-logs\") pod \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.963102 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-public-tls-certs\") pod \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.963209 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.963237 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-httpd-run\") pod \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.963260 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-combined-ca-bundle\") pod \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.963302 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-scripts\") pod \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.963327 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slzm5\" (UniqueName: \"kubernetes.io/projected/be14f196-9b19-49e3-bc07-b642c9ae7ff8-kube-api-access-slzm5\") pod \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.967421 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-logs" (OuterVolumeSpecName: "logs") pod "be14f196-9b19-49e3-bc07-b642c9ae7ff8" (UID: "be14f196-9b19-49e3-bc07-b642c9ae7ff8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.968318 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "be14f196-9b19-49e3-bc07-b642c9ae7ff8" (UID: "be14f196-9b19-49e3-bc07-b642c9ae7ff8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.974425 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be14f196-9b19-49e3-bc07-b642c9ae7ff8-kube-api-access-slzm5" (OuterVolumeSpecName: "kube-api-access-slzm5") pod "be14f196-9b19-49e3-bc07-b642c9ae7ff8" (UID: "be14f196-9b19-49e3-bc07-b642c9ae7ff8"). InnerVolumeSpecName "kube-api-access-slzm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.975797 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-scripts" (OuterVolumeSpecName: "scripts") pod "be14f196-9b19-49e3-bc07-b642c9ae7ff8" (UID: "be14f196-9b19-49e3-bc07-b642c9ae7ff8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:05 crc kubenswrapper[4936]: I0320 16:21:05.979752 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "be14f196-9b19-49e3-bc07-b642c9ae7ff8" (UID: "be14f196-9b19-49e3-bc07-b642c9ae7ff8"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.028428 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be14f196-9b19-49e3-bc07-b642c9ae7ff8" (UID: "be14f196-9b19-49e3-bc07-b642c9ae7ff8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.060405 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-config-data" (OuterVolumeSpecName: "config-data") pod "be14f196-9b19-49e3-bc07-b642c9ae7ff8" (UID: "be14f196-9b19-49e3-bc07-b642c9ae7ff8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.065226 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be14f196-9b19-49e3-bc07-b642c9ae7ff8" (UID: "be14f196-9b19-49e3-bc07-b642c9ae7ff8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.065291 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-public-tls-certs\") pod \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\" (UID: \"be14f196-9b19-49e3-bc07-b642c9ae7ff8\") " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.065885 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.065909 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.065938 4936 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.065953 4936 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be14f196-9b19-49e3-bc07-b642c9ae7ff8-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.065965 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.065978 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.066001 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slzm5\" (UniqueName: \"kubernetes.io/projected/be14f196-9b19-49e3-bc07-b642c9ae7ff8-kube-api-access-slzm5\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: W0320 16:21:06.065952 4936 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/be14f196-9b19-49e3-bc07-b642c9ae7ff8/volumes/kubernetes.io~secret/public-tls-certs Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.066026 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be14f196-9b19-49e3-bc07-b642c9ae7ff8" (UID: "be14f196-9b19-49e3-bc07-b642c9ae7ff8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.088820 4936 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.088853 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.168239 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-logs\") pod \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.168312 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-httpd-run\") pod \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.168344 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb8bs\" (UniqueName: \"kubernetes.io/projected/601c14d9-f6a4-4f7c-b033-7a35d41594a2-kube-api-access-pb8bs\") pod \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.168434 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-config-data\") pod \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.168466 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-internal-tls-certs\") pod \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.168658 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-combined-ca-bundle\") pod \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.168709 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-scripts\") pod \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.168771 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\" (UID: \"601c14d9-f6a4-4f7c-b033-7a35d41594a2\") " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.168976 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-logs" (OuterVolumeSpecName: "logs") pod "601c14d9-f6a4-4f7c-b033-7a35d41594a2" (UID: "601c14d9-f6a4-4f7c-b033-7a35d41594a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.169823 4936 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be14f196-9b19-49e3-bc07-b642c9ae7ff8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.169852 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.169893 4936 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.172221 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "601c14d9-f6a4-4f7c-b033-7a35d41594a2" (UID: "601c14d9-f6a4-4f7c-b033-7a35d41594a2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.172694 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "601c14d9-f6a4-4f7c-b033-7a35d41594a2" (UID: "601c14d9-f6a4-4f7c-b033-7a35d41594a2"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.173258 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/601c14d9-f6a4-4f7c-b033-7a35d41594a2-kube-api-access-pb8bs" (OuterVolumeSpecName: "kube-api-access-pb8bs") pod "601c14d9-f6a4-4f7c-b033-7a35d41594a2" (UID: "601c14d9-f6a4-4f7c-b033-7a35d41594a2"). InnerVolumeSpecName "kube-api-access-pb8bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.175092 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-scripts" (OuterVolumeSpecName: "scripts") pod "601c14d9-f6a4-4f7c-b033-7a35d41594a2" (UID: "601c14d9-f6a4-4f7c-b033-7a35d41594a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.196706 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "601c14d9-f6a4-4f7c-b033-7a35d41594a2" (UID: "601c14d9-f6a4-4f7c-b033-7a35d41594a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.226618 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "601c14d9-f6a4-4f7c-b033-7a35d41594a2" (UID: "601c14d9-f6a4-4f7c-b033-7a35d41594a2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.232200 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-config-data" (OuterVolumeSpecName: "config-data") pod "601c14d9-f6a4-4f7c-b033-7a35d41594a2" (UID: "601c14d9-f6a4-4f7c-b033-7a35d41594a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.271224 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.271265 4936 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.271276 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.271284 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601c14d9-f6a4-4f7c-b033-7a35d41594a2-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.271316 4936 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.271325 4936 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/601c14d9-f6a4-4f7c-b033-7a35d41594a2-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.271333 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb8bs\" (UniqueName: \"kubernetes.io/projected/601c14d9-f6a4-4f7c-b033-7a35d41594a2-kube-api-access-pb8bs\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.288886 4936 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.317432 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.373256 4936 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.653354 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerStarted","Data":"e8dd8fd20ecc12cd9f023b882cfc64bc65a2011c982f9eba9b9c664cd28efb57"} Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.655054 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.655047 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"601c14d9-f6a4-4f7c-b033-7a35d41594a2","Type":"ContainerDied","Data":"8601f72ec3acf0b1e58ec2f8002d449a7e5235cf88ca804977b7ac869c54f6f0"} Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.655188 4936 scope.go:117] "RemoveContainer" containerID="b4d22dfeb13ddda6b999a4c43aab3870c668690858a2a94abfcfdc556416edd3" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.673896 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hspgq" event={"ID":"49bf6254-7f19-41f0-969b-fc8b338ece00","Type":"ContainerStarted","Data":"df2185c05f090644e7cfe6108e193ef3f7821f9aa7a7ebb86bba8d2c486395aa"} Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.683905 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be14f196-9b19-49e3-bc07-b642c9ae7ff8","Type":"ContainerDied","Data":"0117ecdb8c9b0d9c85684fa1f18277bfb3211cc39c579b96b4ea9282d9bf8bbe"} Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.684050 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.692635 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.702000 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.710422 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:21:06 crc kubenswrapper[4936]: E0320 16:21:06.710811 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerName="glance-httpd" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.710830 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerName="glance-httpd" Mar 20 16:21:06 crc kubenswrapper[4936]: E0320 16:21:06.710842 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerName="glance-httpd" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.710849 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerName="glance-httpd" Mar 20 16:21:06 crc kubenswrapper[4936]: E0320 16:21:06.710887 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerName="glance-log" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.710896 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerName="glance-log" Mar 20 16:21:06 crc kubenswrapper[4936]: E0320 16:21:06.710908 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerName="glance-log" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.710914 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerName="glance-log" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.711061 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerName="glance-log" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.711080 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerName="glance-log" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.711092 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" containerName="glance-httpd" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.711098 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" containerName="glance-httpd" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.711757 4936 scope.go:117] "RemoveContainer" containerID="5d07190ffa6702b331b01eafb1e8f1975d312c0eb6d60b7f88fce924ab3951af" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.711980 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.715073 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.715095 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.715337 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-l2w6m" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.715998 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.720164 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-hspgq" podStartSLOduration=1.922982525 podStartE2EDuration="11.720141403s" podCreationTimestamp="2026-03-20 16:20:55 +0000 UTC" firstStartedPulling="2026-03-20 16:20:56.083627144 +0000 UTC m=+1207.029994969" lastFinishedPulling="2026-03-20 16:21:05.880786032 +0000 UTC m=+1216.827153847" observedRunningTime="2026-03-20 16:21:06.707980996 +0000 UTC m=+1217.654348811" watchObservedRunningTime="2026-03-20 16:21:06.720141403 +0000 UTC m=+1217.666509228" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.738267 4936 scope.go:117] "RemoveContainer" containerID="1ca8e48a473928080395ecdceb6e5c0280a91455aee8c60600269c60c57486d4" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.740632 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.780356 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3640384f-37b8-43b3-8fe2-10359b9482a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.780415 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.780475 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.780494 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.780517 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.781501 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3640384f-37b8-43b3-8fe2-10359b9482a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.781524 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ndmd\" (UniqueName: \"kubernetes.io/projected/3640384f-37b8-43b3-8fe2-10359b9482a8-kube-api-access-9ndmd\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.781560 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.787939 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.802989 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.803773 4936 scope.go:117] "RemoveContainer" containerID="9433d1ec221bb76adc3c454f903b2634b9e3a2292e74d1bbd6adb254492f958d" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.822707 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.824123 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.825937 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.826125 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.834276 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.883020 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.883106 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3640384f-37b8-43b3-8fe2-10359b9482a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.883135 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.883189 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.883203 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.883233 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.883264 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3640384f-37b8-43b3-8fe2-10359b9482a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.883286 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ndmd\" (UniqueName: \"kubernetes.io/projected/3640384f-37b8-43b3-8fe2-10359b9482a8-kube-api-access-9ndmd\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.883703 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.884113 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3640384f-37b8-43b3-8fe2-10359b9482a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.885017 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3640384f-37b8-43b3-8fe2-10359b9482a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.889208 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.889231 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.892493 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.892924 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3640384f-37b8-43b3-8fe2-10359b9482a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.904867 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ndmd\" (UniqueName: \"kubernetes.io/projected/3640384f-37b8-43b3-8fe2-10359b9482a8-kube-api-access-9ndmd\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.921029 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"3640384f-37b8-43b3-8fe2-10359b9482a8\") " pod="openstack/glance-default-internal-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.985286 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.985342 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.985474 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-logs\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.985593 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.985656 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.985690 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.985925 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt8qf\" (UniqueName: \"kubernetes.io/projected/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-kube-api-access-lt8qf\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:06 crc kubenswrapper[4936]: I0320 16:21:06.986010 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.033523 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.087902 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.087957 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.087977 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.088048 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-logs\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.088103 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.088126 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.088144 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.088187 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt8qf\" (UniqueName: \"kubernetes.io/projected/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-kube-api-access-lt8qf\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.089275 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-logs\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.089440 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.089758 4936 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.092240 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.092913 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.092922 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.112512 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.118348 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt8qf\" (UniqueName: \"kubernetes.io/projected/dc2d8076-f00b-4d31-89ba-4f15a9d6d50a-kube-api-access-lt8qf\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.123678 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a\") " pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.157791 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.698168 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerStarted","Data":"3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997"} Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.734333 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.837481 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 20 16:21:07 crc kubenswrapper[4936]: W0320 16:21:07.845093 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3640384f_37b8_43b3_8fe2_10359b9482a8.slice/crio-f2917282425323ba8e2da60b7ef78199dda09aeb67d4cb449ee1f6f4a2003bf1 WatchSource:0}: Error finding container f2917282425323ba8e2da60b7ef78199dda09aeb67d4cb449ee1f6f4a2003bf1: Status 404 returned error can't find the container with id f2917282425323ba8e2da60b7ef78199dda09aeb67d4cb449ee1f6f4a2003bf1 Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.873306 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="601c14d9-f6a4-4f7c-b033-7a35d41594a2" path="/var/lib/kubelet/pods/601c14d9-f6a4-4f7c-b033-7a35d41594a2/volumes" Mar 20 16:21:07 crc kubenswrapper[4936]: I0320 16:21:07.874328 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be14f196-9b19-49e3-bc07-b642c9ae7ff8" path="/var/lib/kubelet/pods/be14f196-9b19-49e3-bc07-b642c9ae7ff8/volumes" Mar 20 16:21:08 crc kubenswrapper[4936]: I0320 16:21:08.719843 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerStarted","Data":"e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d"} Mar 20 16:21:08 crc kubenswrapper[4936]: I0320 16:21:08.722079 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3640384f-37b8-43b3-8fe2-10359b9482a8","Type":"ContainerStarted","Data":"bdfa55bc3bad1c63a36b75f20556423ea1e99356ea913ffa98527d7b99708564"} Mar 20 16:21:08 crc kubenswrapper[4936]: I0320 16:21:08.722104 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3640384f-37b8-43b3-8fe2-10359b9482a8","Type":"ContainerStarted","Data":"f2917282425323ba8e2da60b7ef78199dda09aeb67d4cb449ee1f6f4a2003bf1"} Mar 20 16:21:08 crc kubenswrapper[4936]: I0320 16:21:08.723755 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a","Type":"ContainerStarted","Data":"465152bb037b2b4f0736924d35bc3290cb475d662f6b8a8c3331726aae2f4657"} Mar 20 16:21:08 crc kubenswrapper[4936]: I0320 16:21:08.723779 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a","Type":"ContainerStarted","Data":"767cfb1c6dbc32103f9d35223b448a6d9486b56cf9b9a748b9b47a6f85f9ad43"} Mar 20 16:21:09 crc kubenswrapper[4936]: I0320 16:21:09.736013 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerStarted","Data":"6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041"} Mar 20 16:21:09 crc kubenswrapper[4936]: I0320 16:21:09.737821 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3640384f-37b8-43b3-8fe2-10359b9482a8","Type":"ContainerStarted","Data":"ddfe75efafffd0b3ff34506630ad3d0705a92376f97c5ca334244c7b2dd2b4ee"} Mar 20 16:21:09 crc kubenswrapper[4936]: I0320 16:21:09.739594 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc2d8076-f00b-4d31-89ba-4f15a9d6d50a","Type":"ContainerStarted","Data":"a4181823a80f5c5e750c926fc19d4d999c24c0dfbf3cc65d389ab5584cbab7d6"} Mar 20 16:21:09 crc kubenswrapper[4936]: I0320 16:21:09.760186 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.7601626870000002 podStartE2EDuration="3.760162687s" podCreationTimestamp="2026-03-20 16:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:09.755049575 +0000 UTC m=+1220.701417400" watchObservedRunningTime="2026-03-20 16:21:09.760162687 +0000 UTC m=+1220.706530502" Mar 20 16:21:09 crc kubenswrapper[4936]: I0320 16:21:09.786884 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.7868585059999997 podStartE2EDuration="3.786858506s" podCreationTimestamp="2026-03-20 16:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:09.779477352 +0000 UTC m=+1220.725845167" watchObservedRunningTime="2026-03-20 16:21:09.786858506 +0000 UTC m=+1220.733226321" Mar 20 16:21:11 crc kubenswrapper[4936]: I0320 16:21:11.756154 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerStarted","Data":"69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac"} Mar 20 16:21:11 crc kubenswrapper[4936]: I0320 16:21:11.756677 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 20 16:21:11 crc kubenswrapper[4936]: I0320 16:21:11.756495 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="ceilometer-notification-agent" containerID="cri-o://e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d" gracePeriod=30 Mar 20 16:21:11 crc kubenswrapper[4936]: I0320 16:21:11.756349 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="ceilometer-central-agent" containerID="cri-o://3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997" gracePeriod=30 Mar 20 16:21:11 crc kubenswrapper[4936]: I0320 16:21:11.756467 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="sg-core" containerID="cri-o://6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041" gracePeriod=30 Mar 20 16:21:11 crc kubenswrapper[4936]: I0320 16:21:11.756528 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="proxy-httpd" containerID="cri-o://69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac" gracePeriod=30 Mar 20 16:21:11 crc kubenswrapper[4936]: I0320 16:21:11.786079 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.166554064 podStartE2EDuration="12.786057141s" podCreationTimestamp="2026-03-20 16:20:59 +0000 UTC" firstStartedPulling="2026-03-20 16:21:06.321445744 +0000 UTC m=+1217.267813559" lastFinishedPulling="2026-03-20 16:21:10.940948821 +0000 UTC m=+1221.887316636" observedRunningTime="2026-03-20 16:21:11.780427425 +0000 UTC m=+1222.726795260" watchObservedRunningTime="2026-03-20 16:21:11.786057141 +0000 UTC m=+1222.732424956" Mar 20 16:21:12 crc kubenswrapper[4936]: I0320 16:21:12.769449 4936 generic.go:334] "Generic (PLEG): container finished" podID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerID="69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac" exitCode=0 Mar 20 16:21:12 crc kubenswrapper[4936]: I0320 16:21:12.769775 4936 generic.go:334] "Generic (PLEG): container finished" podID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerID="6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041" exitCode=2 Mar 20 16:21:12 crc kubenswrapper[4936]: I0320 16:21:12.769787 4936 generic.go:334] "Generic (PLEG): container finished" podID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerID="e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d" exitCode=0 Mar 20 16:21:12 crc kubenswrapper[4936]: I0320 16:21:12.769527 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerDied","Data":"69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac"} Mar 20 16:21:12 crc kubenswrapper[4936]: I0320 16:21:12.769836 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerDied","Data":"6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041"} Mar 20 16:21:12 crc kubenswrapper[4936]: I0320 16:21:12.769855 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerDied","Data":"e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d"} Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.249994 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.327763 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-run-httpd\") pod \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.327875 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-scripts\") pod \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.328006 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-log-httpd\") pod \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.328073 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44wl9\" (UniqueName: \"kubernetes.io/projected/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-kube-api-access-44wl9\") pod \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.328129 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-sg-core-conf-yaml\") pod \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.328175 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" (UID: "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.328262 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-config-data\") pod \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.328326 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-combined-ca-bundle\") pod \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\" (UID: \"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0\") " Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.328527 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" (UID: "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.329398 4936 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.329427 4936 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.332860 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-kube-api-access-44wl9" (OuterVolumeSpecName: "kube-api-access-44wl9") pod "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" (UID: "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0"). InnerVolumeSpecName "kube-api-access-44wl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.333339 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-scripts" (OuterVolumeSpecName: "scripts") pod "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" (UID: "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.355732 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" (UID: "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.403486 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" (UID: "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.424050 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-config-data" (OuterVolumeSpecName: "config-data") pod "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" (UID: "4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.431186 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.431219 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44wl9\" (UniqueName: \"kubernetes.io/projected/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-kube-api-access-44wl9\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.431230 4936 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.431239 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.431248 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.788570 4936 generic.go:334] "Generic (PLEG): container finished" podID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerID="3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997" exitCode=0 Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.788612 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerDied","Data":"3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997"} Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.788638 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0","Type":"ContainerDied","Data":"e8dd8fd20ecc12cd9f023b882cfc64bc65a2011c982f9eba9b9c664cd28efb57"} Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.788638 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.788654 4936 scope.go:117] "RemoveContainer" containerID="69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.816659 4936 scope.go:117] "RemoveContainer" containerID="6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.822793 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.833053 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.848351 4936 scope.go:117] "RemoveContainer" containerID="e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.849733 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:14 crc kubenswrapper[4936]: E0320 16:21:14.850190 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="ceilometer-notification-agent" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.850218 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="ceilometer-notification-agent" Mar 20 16:21:14 crc kubenswrapper[4936]: E0320 16:21:14.850237 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="proxy-httpd" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.850245 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="proxy-httpd" Mar 20 16:21:14 crc kubenswrapper[4936]: E0320 16:21:14.850263 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="ceilometer-central-agent" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.850271 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="ceilometer-central-agent" Mar 20 16:21:14 crc kubenswrapper[4936]: E0320 16:21:14.850298 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="sg-core" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.850307 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="sg-core" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.850519 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="ceilometer-central-agent" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.850554 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="ceilometer-notification-agent" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.850579 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="proxy-httpd" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.850589 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" containerName="sg-core" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.855496 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.858113 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.858690 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.898904 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.914288 4936 scope.go:117] "RemoveContainer" containerID="3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.939430 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-scripts\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.939472 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-log-httpd\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.939519 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-config-data\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.939569 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.939619 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.939651 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qdg7\" (UniqueName: \"kubernetes.io/projected/cf000d82-6986-4c78-80bf-7913eeed68b9-kube-api-access-6qdg7\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.939785 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-run-httpd\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.981016 4936 scope.go:117] "RemoveContainer" containerID="69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac" Mar 20 16:21:14 crc kubenswrapper[4936]: E0320 16:21:14.981518 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac\": container with ID starting with 69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac not found: ID does not exist" containerID="69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.981685 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac"} err="failed to get container status \"69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac\": rpc error: code = NotFound desc = could not find container \"69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac\": container with ID starting with 69da74015ed6486c8c70b3bc6bec60c4199be7046e755984ac3d88a34c875dac not found: ID does not exist" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.981770 4936 scope.go:117] "RemoveContainer" containerID="6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041" Mar 20 16:21:14 crc kubenswrapper[4936]: E0320 16:21:14.982141 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041\": container with ID starting with 6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041 not found: ID does not exist" containerID="6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.982190 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041"} err="failed to get container status \"6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041\": rpc error: code = NotFound desc = could not find container \"6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041\": container with ID starting with 6879a30c87751d47f6293e077bfa6a97ba54216914faf8ec96f9fb69bd4bc041 not found: ID does not exist" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.982219 4936 scope.go:117] "RemoveContainer" containerID="e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d" Mar 20 16:21:14 crc kubenswrapper[4936]: E0320 16:21:14.982683 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d\": container with ID starting with e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d not found: ID does not exist" containerID="e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.982730 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d"} err="failed to get container status \"e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d\": rpc error: code = NotFound desc = could not find container \"e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d\": container with ID starting with e2568dba46d0e83e3408081ebe01b452f966fc7f06a00fd7d986b4aeb304167d not found: ID does not exist" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.982744 4936 scope.go:117] "RemoveContainer" containerID="3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997" Mar 20 16:21:14 crc kubenswrapper[4936]: E0320 16:21:14.983016 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997\": container with ID starting with 3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997 not found: ID does not exist" containerID="3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997" Mar 20 16:21:14 crc kubenswrapper[4936]: I0320 16:21:14.983122 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997"} err="failed to get container status \"3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997\": rpc error: code = NotFound desc = could not find container \"3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997\": container with ID starting with 3b93ea3f8eebb7630eddd772abcff2eb9384288f651fd0c698efb163b202a997 not found: ID does not exist" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.041577 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-run-httpd\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.041666 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-scripts\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.041689 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-log-httpd\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.041735 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-config-data\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.041759 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.041829 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.041856 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qdg7\" (UniqueName: \"kubernetes.io/projected/cf000d82-6986-4c78-80bf-7913eeed68b9-kube-api-access-6qdg7\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.042847 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-log-httpd\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.043120 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-run-httpd\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.047493 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-scripts\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.047671 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.047711 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-config-data\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.061352 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qdg7\" (UniqueName: \"kubernetes.io/projected/cf000d82-6986-4c78-80bf-7913eeed68b9-kube-api-access-6qdg7\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.061515 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.176090 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:21:15 crc kubenswrapper[4936]: W0320 16:21:15.612369 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf000d82_6986_4c78_80bf_7913eeed68b9.slice/crio-5d3975c8ea69b0efa6886971825c8aa23fede17c8c0fff103c1e8b534dc95ca2 WatchSource:0}: Error finding container 5d3975c8ea69b0efa6886971825c8aa23fede17c8c0fff103c1e8b534dc95ca2: Status 404 returned error can't find the container with id 5d3975c8ea69b0efa6886971825c8aa23fede17c8c0fff103c1e8b534dc95ca2 Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.614579 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.800164 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerStarted","Data":"5d3975c8ea69b0efa6886971825c8aa23fede17c8c0fff103c1e8b534dc95ca2"} Mar 20 16:21:15 crc kubenswrapper[4936]: I0320 16:21:15.866097 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0" path="/var/lib/kubelet/pods/4b412d3c-2e1e-4dd8-b4d3-0712b2f6dbb0/volumes" Mar 20 16:21:16 crc kubenswrapper[4936]: I0320 16:21:16.812516 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerStarted","Data":"5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b"} Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.034695 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.034762 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.071518 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.080978 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.158227 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.158604 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.201297 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.215754 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.823781 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerStarted","Data":"59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767"} Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.824598 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.824641 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.824654 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 20 16:21:17 crc kubenswrapper[4936]: I0320 16:21:17.824665 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 20 16:21:18 crc kubenswrapper[4936]: I0320 16:21:18.835231 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerStarted","Data":"eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031"} Mar 20 16:21:19 crc kubenswrapper[4936]: I0320 16:21:19.846324 4936 generic.go:334] "Generic (PLEG): container finished" podID="49bf6254-7f19-41f0-969b-fc8b338ece00" containerID="df2185c05f090644e7cfe6108e193ef3f7821f9aa7a7ebb86bba8d2c486395aa" exitCode=0 Mar 20 16:21:19 crc kubenswrapper[4936]: I0320 16:21:19.846400 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hspgq" event={"ID":"49bf6254-7f19-41f0-969b-fc8b338ece00","Type":"ContainerDied","Data":"df2185c05f090644e7cfe6108e193ef3f7821f9aa7a7ebb86bba8d2c486395aa"} Mar 20 16:21:19 crc kubenswrapper[4936]: I0320 16:21:19.890525 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:19 crc kubenswrapper[4936]: I0320 16:21:19.890981 4936 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 16:21:19 crc kubenswrapper[4936]: I0320 16:21:19.937827 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 20 16:21:20 crc kubenswrapper[4936]: I0320 16:21:20.036612 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 20 16:21:20 crc kubenswrapper[4936]: I0320 16:21:20.036736 4936 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 16:21:20 crc kubenswrapper[4936]: I0320 16:21:20.039640 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 20 16:21:20 crc kubenswrapper[4936]: I0320 16:21:20.856856 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerStarted","Data":"6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4"} Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.282850 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.310786 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.80215764 podStartE2EDuration="7.310755476s" podCreationTimestamp="2026-03-20 16:21:14 +0000 UTC" firstStartedPulling="2026-03-20 16:21:15.614828494 +0000 UTC m=+1226.561196299" lastFinishedPulling="2026-03-20 16:21:20.12342632 +0000 UTC m=+1231.069794135" observedRunningTime="2026-03-20 16:21:20.887662641 +0000 UTC m=+1231.834030466" watchObservedRunningTime="2026-03-20 16:21:21.310755476 +0000 UTC m=+1232.257123321" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.464835 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-config-data\") pod \"49bf6254-7f19-41f0-969b-fc8b338ece00\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.464985 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-combined-ca-bundle\") pod \"49bf6254-7f19-41f0-969b-fc8b338ece00\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.465022 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2djg\" (UniqueName: \"kubernetes.io/projected/49bf6254-7f19-41f0-969b-fc8b338ece00-kube-api-access-q2djg\") pod \"49bf6254-7f19-41f0-969b-fc8b338ece00\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.465057 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-scripts\") pod \"49bf6254-7f19-41f0-969b-fc8b338ece00\" (UID: \"49bf6254-7f19-41f0-969b-fc8b338ece00\") " Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.473191 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49bf6254-7f19-41f0-969b-fc8b338ece00-kube-api-access-q2djg" (OuterVolumeSpecName: "kube-api-access-q2djg") pod "49bf6254-7f19-41f0-969b-fc8b338ece00" (UID: "49bf6254-7f19-41f0-969b-fc8b338ece00"). InnerVolumeSpecName "kube-api-access-q2djg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.473087 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-scripts" (OuterVolumeSpecName: "scripts") pod "49bf6254-7f19-41f0-969b-fc8b338ece00" (UID: "49bf6254-7f19-41f0-969b-fc8b338ece00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.500719 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-config-data" (OuterVolumeSpecName: "config-data") pod "49bf6254-7f19-41f0-969b-fc8b338ece00" (UID: "49bf6254-7f19-41f0-969b-fc8b338ece00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.504199 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49bf6254-7f19-41f0-969b-fc8b338ece00" (UID: "49bf6254-7f19-41f0-969b-fc8b338ece00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.566747 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.566781 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.566793 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2djg\" (UniqueName: \"kubernetes.io/projected/49bf6254-7f19-41f0-969b-fc8b338ece00-kube-api-access-q2djg\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.566801 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49bf6254-7f19-41f0-969b-fc8b338ece00-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.885309 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hspgq" event={"ID":"49bf6254-7f19-41f0-969b-fc8b338ece00","Type":"ContainerDied","Data":"e777ea75fe121c555066848cd2b59a9d28426b9cc6c251ba6c3ef4f656cbaf15"} Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.885632 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e777ea75fe121c555066848cd2b59a9d28426b9cc6c251ba6c3ef4f656cbaf15" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.885658 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.885349 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hspgq" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.982147 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 20 16:21:21 crc kubenswrapper[4936]: E0320 16:21:21.982580 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bf6254-7f19-41f0-969b-fc8b338ece00" containerName="nova-cell0-conductor-db-sync" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.982596 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bf6254-7f19-41f0-969b-fc8b338ece00" containerName="nova-cell0-conductor-db-sync" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.982777 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="49bf6254-7f19-41f0-969b-fc8b338ece00" containerName="nova-cell0-conductor-db-sync" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.983473 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.991928 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-jmd6p" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.992134 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 20 16:21:21 crc kubenswrapper[4936]: I0320 16:21:21.993269 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.075446 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f2d11d-47b6-4791-88c9-25911a5b6bc7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d4f2d11d-47b6-4791-88c9-25911a5b6bc7\") " pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.075506 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmmd8\" (UniqueName: \"kubernetes.io/projected/d4f2d11d-47b6-4791-88c9-25911a5b6bc7-kube-api-access-rmmd8\") pod \"nova-cell0-conductor-0\" (UID: \"d4f2d11d-47b6-4791-88c9-25911a5b6bc7\") " pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.075575 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f2d11d-47b6-4791-88c9-25911a5b6bc7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d4f2d11d-47b6-4791-88c9-25911a5b6bc7\") " pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.176954 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f2d11d-47b6-4791-88c9-25911a5b6bc7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d4f2d11d-47b6-4791-88c9-25911a5b6bc7\") " pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.177047 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmmd8\" (UniqueName: \"kubernetes.io/projected/d4f2d11d-47b6-4791-88c9-25911a5b6bc7-kube-api-access-rmmd8\") pod \"nova-cell0-conductor-0\" (UID: \"d4f2d11d-47b6-4791-88c9-25911a5b6bc7\") " pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.177091 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f2d11d-47b6-4791-88c9-25911a5b6bc7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d4f2d11d-47b6-4791-88c9-25911a5b6bc7\") " pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.185341 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f2d11d-47b6-4791-88c9-25911a5b6bc7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d4f2d11d-47b6-4791-88c9-25911a5b6bc7\") " pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.186345 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f2d11d-47b6-4791-88c9-25911a5b6bc7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d4f2d11d-47b6-4791-88c9-25911a5b6bc7\") " pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.199750 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmmd8\" (UniqueName: \"kubernetes.io/projected/d4f2d11d-47b6-4791-88c9-25911a5b6bc7-kube-api-access-rmmd8\") pod \"nova-cell0-conductor-0\" (UID: \"d4f2d11d-47b6-4791-88c9-25911a5b6bc7\") " pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.320039 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.787377 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 20 16:21:22 crc kubenswrapper[4936]: I0320 16:21:22.895252 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d4f2d11d-47b6-4791-88c9-25911a5b6bc7","Type":"ContainerStarted","Data":"9fe9c2cc17f97b98366def2e996de41af567a5e1462f02f8b68a7684672c4484"} Mar 20 16:21:23 crc kubenswrapper[4936]: I0320 16:21:23.904853 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d4f2d11d-47b6-4791-88c9-25911a5b6bc7","Type":"ContainerStarted","Data":"5b3b6f1a91578ff77c7c239ce99a1eac0309a71e9533c8f57ad01f7480859006"} Mar 20 16:21:23 crc kubenswrapper[4936]: I0320 16:21:23.905138 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:23 crc kubenswrapper[4936]: I0320 16:21:23.923584 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.923561006 podStartE2EDuration="2.923561006s" podCreationTimestamp="2026-03-20 16:21:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:23.921460046 +0000 UTC m=+1234.867827881" watchObservedRunningTime="2026-03-20 16:21:23.923561006 +0000 UTC m=+1234.869928811" Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.349246 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.811003 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-k54pt"] Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.812443 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.815920 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.816128 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.823972 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-k54pt"] Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.999389 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.999556 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-scripts\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.999642 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-config-data\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:27 crc kubenswrapper[4936]: I0320 16:21:27.999666 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8s79\" (UniqueName: \"kubernetes.io/projected/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-kube-api-access-z8s79\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.088616 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.103307 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-config-data\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.103370 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8s79\" (UniqueName: \"kubernetes.io/projected/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-kube-api-access-z8s79\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.103471 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.103577 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-scripts\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.107535 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.139669 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.140094 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-scripts\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.140329 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-config-data\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.147392 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.149709 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8s79\" (UniqueName: \"kubernetes.io/projected/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-kube-api-access-z8s79\") pod \"nova-cell0-cell-mapping-k54pt\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.157023 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.159231 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.178401 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.180961 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.208843 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.209774 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.210071 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjdq7\" (UniqueName: \"kubernetes.io/projected/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-kube-api-access-qjdq7\") pod \"nova-cell1-novncproxy-0\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.238154 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.264709 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.266390 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.274829 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.308617 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.311722 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjdq7\" (UniqueName: \"kubernetes.io/projected/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-kube-api-access-qjdq7\") pod \"nova-cell1-novncproxy-0\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.311787 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-config-data\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.311820 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.311849 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b930af4-f185-44c9-a531-b0313021e3a1-logs\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.311869 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj77n\" (UniqueName: \"kubernetes.io/projected/4b930af4-f185-44c9-a531-b0313021e3a1-kube-api-access-nj77n\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.311905 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.311965 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.320355 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.329747 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.333105 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.334143 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjdq7\" (UniqueName: \"kubernetes.io/projected/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-kube-api-access-qjdq7\") pod \"nova-cell1-novncproxy-0\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.336336 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.338261 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.384281 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.420793 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-8b4pt"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.422736 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.423195 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-config-data\") pod \"nova-scheduler-0\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.423330 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-logs\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.423390 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-config-data\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.423462 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7flt\" (UniqueName: \"kubernetes.io/projected/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-kube-api-access-z7flt\") pod \"nova-scheduler-0\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.423503 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b930af4-f185-44c9-a531-b0313021e3a1-logs\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.423521 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.423590 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj77n\" (UniqueName: \"kubernetes.io/projected/4b930af4-f185-44c9-a531-b0313021e3a1-kube-api-access-nj77n\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.423645 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.423734 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-config-data\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.424200 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b930af4-f185-44c9-a531-b0313021e3a1-logs\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.424333 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpldg\" (UniqueName: \"kubernetes.io/projected/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-kube-api-access-fpldg\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.424358 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.435585 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-config-data\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.437053 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-8b4pt"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.437264 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.443922 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj77n\" (UniqueName: \"kubernetes.io/projected/4b930af4-f185-44c9-a531-b0313021e3a1-kube-api-access-nj77n\") pod \"nova-metadata-0\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.444359 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.527430 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.527559 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-logs\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.527635 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-svc\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.527662 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pzjs\" (UniqueName: \"kubernetes.io/projected/a481e8ea-35e6-4ad7-8fea-43226787831b-kube-api-access-2pzjs\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.527698 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7flt\" (UniqueName: \"kubernetes.io/projected/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-kube-api-access-z7flt\") pod \"nova-scheduler-0\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.527721 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-config\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.527743 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.528194 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-logs\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.528881 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.529475 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.529507 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.529823 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-config-data\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.529898 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpldg\" (UniqueName: \"kubernetes.io/projected/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-kube-api-access-fpldg\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.530273 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-config-data\") pod \"nova-scheduler-0\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.532906 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.533652 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.534098 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-config-data\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.534205 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-config-data\") pod \"nova-scheduler-0\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.546174 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.554606 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7flt\" (UniqueName: \"kubernetes.io/projected/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-kube-api-access-z7flt\") pod \"nova-scheduler-0\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.554755 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpldg\" (UniqueName: \"kubernetes.io/projected/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-kube-api-access-fpldg\") pod \"nova-api-0\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.565809 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.596129 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.632489 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.632572 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-svc\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.632603 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pzjs\" (UniqueName: \"kubernetes.io/projected/a481e8ea-35e6-4ad7-8fea-43226787831b-kube-api-access-2pzjs\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.632645 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-config\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.632692 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.632715 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.635281 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.635935 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-svc\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.637772 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-config\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.638347 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.638895 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.697337 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pzjs\" (UniqueName: \"kubernetes.io/projected/a481e8ea-35e6-4ad7-8fea-43226787831b-kube-api-access-2pzjs\") pod \"dnsmasq-dns-757b4f8459-8b4pt\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.796207 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.811315 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.923471 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fvb75"] Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.926944 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.929630 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.931646 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 20 16:21:28 crc kubenswrapper[4936]: I0320 16:21:28.956120 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fvb75"] Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.045694 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.045757 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-config-data\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.045790 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jf8x\" (UniqueName: \"kubernetes.io/projected/6071aca2-12af-427f-88a0-5661b147f6d8-kube-api-access-7jf8x\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.045893 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-scripts\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.110144 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-k54pt"] Mar 20 16:21:29 crc kubenswrapper[4936]: W0320 16:21:29.124242 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod446d671c_a5e1_4430_9aa6_262bb3a9b2d0.slice/crio-749162c9811e511d80bfc53d7b063a0e0cf5110c719edb86dd2d565e8155f139 WatchSource:0}: Error finding container 749162c9811e511d80bfc53d7b063a0e0cf5110c719edb86dd2d565e8155f139: Status 404 returned error can't find the container with id 749162c9811e511d80bfc53d7b063a0e0cf5110c719edb86dd2d565e8155f139 Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.149749 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.149813 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-config-data\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.149845 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jf8x\" (UniqueName: \"kubernetes.io/projected/6071aca2-12af-427f-88a0-5661b147f6d8-kube-api-access-7jf8x\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.149924 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-scripts\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.155976 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.163083 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-scripts\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.166515 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-config-data\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.173881 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jf8x\" (UniqueName: \"kubernetes.io/projected/6071aca2-12af-427f-88a0-5661b147f6d8-kube-api-access-7jf8x\") pod \"nova-cell1-conductor-db-sync-fvb75\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.254789 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.383852 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.406060 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 20 16:21:29 crc kubenswrapper[4936]: W0320 16:21:29.439496 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99ba5b5a_c9c2_4ec0_b6e8_92802dfd1fcd.slice/crio-4afd3d55b76e422b851f7dca01608c7462a4a35f844417f38c6a33f0094f22ef WatchSource:0}: Error finding container 4afd3d55b76e422b851f7dca01608c7462a4a35f844417f38c6a33f0094f22ef: Status 404 returned error can't find the container with id 4afd3d55b76e422b851f7dca01608c7462a4a35f844417f38c6a33f0094f22ef Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.513947 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:21:29 crc kubenswrapper[4936]: W0320 16:21:29.514998 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74ae4b9f_ac00_44f8_b072_b65bc3bc4373.slice/crio-1ea311a8a00069b5f9b0f06b2034b5bda4937c281a8ccdde2f793cfa7410cb63 WatchSource:0}: Error finding container 1ea311a8a00069b5f9b0f06b2034b5bda4937c281a8ccdde2f793cfa7410cb63: Status 404 returned error can't find the container with id 1ea311a8a00069b5f9b0f06b2034b5bda4937c281a8ccdde2f793cfa7410cb63 Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.576394 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-8b4pt"] Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.594855 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:21:29 crc kubenswrapper[4936]: I0320 16:21:29.839653 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fvb75"] Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.009119 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd","Type":"ContainerStarted","Data":"4afd3d55b76e422b851f7dca01608c7462a4a35f844417f38c6a33f0094f22ef"} Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.011121 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"74ae4b9f-ac00-44f8-b072-b65bc3bc4373","Type":"ContainerStarted","Data":"1ea311a8a00069b5f9b0f06b2034b5bda4937c281a8ccdde2f793cfa7410cb63"} Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.012247 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fvb75" event={"ID":"6071aca2-12af-427f-88a0-5661b147f6d8","Type":"ContainerStarted","Data":"66acb92317c2e5bc49c1cfdc4c7efed8a75d92e13799dd7930dc04127f4e3089"} Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.018243 4936 generic.go:334] "Generic (PLEG): container finished" podID="a481e8ea-35e6-4ad7-8fea-43226787831b" containerID="ad78e2b5994b670e6484af03a854dce490cf809b038545ce04d361a96b4cc354" exitCode=0 Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.018291 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" event={"ID":"a481e8ea-35e6-4ad7-8fea-43226787831b","Type":"ContainerDied","Data":"ad78e2b5994b670e6484af03a854dce490cf809b038545ce04d361a96b4cc354"} Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.018309 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" event={"ID":"a481e8ea-35e6-4ad7-8fea-43226787831b","Type":"ContainerStarted","Data":"491708855fa1d4410b7655b256d92ae9a600e6d88173340c99c7f246387d90c4"} Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.030477 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k54pt" event={"ID":"446d671c-a5e1-4430-9aa6-262bb3a9b2d0","Type":"ContainerStarted","Data":"c18db7a3d9576592f70a01ee0599f74b09c962beacbd3cf95e3dfeeaf1d6b493"} Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.030536 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k54pt" event={"ID":"446d671c-a5e1-4430-9aa6-262bb3a9b2d0","Type":"ContainerStarted","Data":"749162c9811e511d80bfc53d7b063a0e0cf5110c719edb86dd2d565e8155f139"} Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.032471 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b930af4-f185-44c9-a531-b0313021e3a1","Type":"ContainerStarted","Data":"1805a0a9e5a962df02289e8d56249f04ead2a5fee8b0e31a188449fc442360b2"} Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.034341 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5","Type":"ContainerStarted","Data":"f5b56938f186cf46b2957f247c5180de46a78dd704c3a4ebd78b8ef3f32ff715"} Mar 20 16:21:30 crc kubenswrapper[4936]: I0320 16:21:30.103317 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-k54pt" podStartSLOduration=3.103297573 podStartE2EDuration="3.103297573s" podCreationTimestamp="2026-03-20 16:21:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:30.090379076 +0000 UTC m=+1241.036746901" watchObservedRunningTime="2026-03-20 16:21:30.103297573 +0000 UTC m=+1241.049665388" Mar 20 16:21:30 crc kubenswrapper[4936]: E0320 16:21:30.130081 4936 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda481e8ea_35e6_4ad7_8fea_43226787831b.slice/crio-ad78e2b5994b670e6484af03a854dce490cf809b038545ce04d361a96b4cc354.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda481e8ea_35e6_4ad7_8fea_43226787831b.slice/crio-conmon-ad78e2b5994b670e6484af03a854dce490cf809b038545ce04d361a96b4cc354.scope\": RecentStats: unable to find data in memory cache]" Mar 20 16:21:31 crc kubenswrapper[4936]: I0320 16:21:31.046494 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fvb75" event={"ID":"6071aca2-12af-427f-88a0-5661b147f6d8","Type":"ContainerStarted","Data":"d07e02914919c0f0a57dcf0f6adfb03c41bf1841fe9fa0a3c9bdbc2026c6cecc"} Mar 20 16:21:31 crc kubenswrapper[4936]: I0320 16:21:31.051483 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" event={"ID":"a481e8ea-35e6-4ad7-8fea-43226787831b","Type":"ContainerStarted","Data":"098e130da0400ff88cfeb0c131f7492751e7cf87364c27359e6498c43b285422"} Mar 20 16:21:31 crc kubenswrapper[4936]: I0320 16:21:31.051753 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:31 crc kubenswrapper[4936]: I0320 16:21:31.063089 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-fvb75" podStartSLOduration=3.063068874 podStartE2EDuration="3.063068874s" podCreationTimestamp="2026-03-20 16:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:31.061282934 +0000 UTC m=+1242.007650769" watchObservedRunningTime="2026-03-20 16:21:31.063068874 +0000 UTC m=+1242.009436689" Mar 20 16:21:31 crc kubenswrapper[4936]: I0320 16:21:31.082815 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" podStartSLOduration=3.082797795 podStartE2EDuration="3.082797795s" podCreationTimestamp="2026-03-20 16:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:31.08049033 +0000 UTC m=+1242.026858145" watchObservedRunningTime="2026-03-20 16:21:31.082797795 +0000 UTC m=+1242.029165600" Mar 20 16:21:32 crc kubenswrapper[4936]: I0320 16:21:32.000949 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:32 crc kubenswrapper[4936]: I0320 16:21:32.031334 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 20 16:21:34 crc kubenswrapper[4936]: I0320 16:21:34.093117 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"74ae4b9f-ac00-44f8-b072-b65bc3bc4373","Type":"ContainerStarted","Data":"c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623"} Mar 20 16:21:34 crc kubenswrapper[4936]: I0320 16:21:34.095020 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b930af4-f185-44c9-a531-b0313021e3a1","Type":"ContainerStarted","Data":"3d791905de8c15ce67c74e8763624bc7e895adc21c973f512c24cbfc8b64aada"} Mar 20 16:21:34 crc kubenswrapper[4936]: I0320 16:21:34.100719 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5","Type":"ContainerStarted","Data":"a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf"} Mar 20 16:21:34 crc kubenswrapper[4936]: I0320 16:21:34.100761 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5","Type":"ContainerStarted","Data":"1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d"} Mar 20 16:21:34 crc kubenswrapper[4936]: I0320 16:21:34.103465 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd","Type":"ContainerStarted","Data":"bb10f7f7dbf8aa243ed767f0471e602deeb4ca721880d72deee1eab1e3da0d82"} Mar 20 16:21:34 crc kubenswrapper[4936]: I0320 16:21:34.103609 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://bb10f7f7dbf8aa243ed767f0471e602deeb4ca721880d72deee1eab1e3da0d82" gracePeriod=30 Mar 20 16:21:34 crc kubenswrapper[4936]: I0320 16:21:34.120441 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.722647568 podStartE2EDuration="6.120422969s" podCreationTimestamp="2026-03-20 16:21:28 +0000 UTC" firstStartedPulling="2026-03-20 16:21:29.516593476 +0000 UTC m=+1240.462961291" lastFinishedPulling="2026-03-20 16:21:32.914368877 +0000 UTC m=+1243.860736692" observedRunningTime="2026-03-20 16:21:34.111041062 +0000 UTC m=+1245.057408877" watchObservedRunningTime="2026-03-20 16:21:34.120422969 +0000 UTC m=+1245.066790784" Mar 20 16:21:34 crc kubenswrapper[4936]: I0320 16:21:34.140397 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.676571689 podStartE2EDuration="7.140368356s" podCreationTimestamp="2026-03-20 16:21:27 +0000 UTC" firstStartedPulling="2026-03-20 16:21:29.449112528 +0000 UTC m=+1240.395480343" lastFinishedPulling="2026-03-20 16:21:32.912909195 +0000 UTC m=+1243.859277010" observedRunningTime="2026-03-20 16:21:34.129823316 +0000 UTC m=+1245.076191131" watchObservedRunningTime="2026-03-20 16:21:34.140368356 +0000 UTC m=+1245.086736171" Mar 20 16:21:34 crc kubenswrapper[4936]: I0320 16:21:34.158800 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.835395812 podStartE2EDuration="6.158778579s" podCreationTimestamp="2026-03-20 16:21:28 +0000 UTC" firstStartedPulling="2026-03-20 16:21:29.612169222 +0000 UTC m=+1240.558537037" lastFinishedPulling="2026-03-20 16:21:32.935551989 +0000 UTC m=+1243.881919804" observedRunningTime="2026-03-20 16:21:34.153005885 +0000 UTC m=+1245.099373720" watchObservedRunningTime="2026-03-20 16:21:34.158778579 +0000 UTC m=+1245.105146394" Mar 20 16:21:35 crc kubenswrapper[4936]: I0320 16:21:35.119233 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b930af4-f185-44c9-a531-b0313021e3a1","Type":"ContainerStarted","Data":"2e9b3c7bbd7dfeab829b80e0e004d8146f9e37ef4f1b06df54a03a1b71b10daa"} Mar 20 16:21:35 crc kubenswrapper[4936]: I0320 16:21:35.120044 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4b930af4-f185-44c9-a531-b0313021e3a1" containerName="nova-metadata-log" containerID="cri-o://3d791905de8c15ce67c74e8763624bc7e895adc21c973f512c24cbfc8b64aada" gracePeriod=30 Mar 20 16:21:35 crc kubenswrapper[4936]: I0320 16:21:35.120160 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4b930af4-f185-44c9-a531-b0313021e3a1" containerName="nova-metadata-metadata" containerID="cri-o://2e9b3c7bbd7dfeab829b80e0e004d8146f9e37ef4f1b06df54a03a1b71b10daa" gracePeriod=30 Mar 20 16:21:35 crc kubenswrapper[4936]: I0320 16:21:35.160239 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.6710294359999995 podStartE2EDuration="8.160218165s" podCreationTimestamp="2026-03-20 16:21:27 +0000 UTC" firstStartedPulling="2026-03-20 16:21:29.429108489 +0000 UTC m=+1240.375476304" lastFinishedPulling="2026-03-20 16:21:32.918297218 +0000 UTC m=+1243.864665033" observedRunningTime="2026-03-20 16:21:35.146722481 +0000 UTC m=+1246.093090296" watchObservedRunningTime="2026-03-20 16:21:35.160218165 +0000 UTC m=+1246.106585980" Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.134850 4936 generic.go:334] "Generic (PLEG): container finished" podID="4b930af4-f185-44c9-a531-b0313021e3a1" containerID="2e9b3c7bbd7dfeab829b80e0e004d8146f9e37ef4f1b06df54a03a1b71b10daa" exitCode=0 Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.135385 4936 generic.go:334] "Generic (PLEG): container finished" podID="4b930af4-f185-44c9-a531-b0313021e3a1" containerID="3d791905de8c15ce67c74e8763624bc7e895adc21c973f512c24cbfc8b64aada" exitCode=143 Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.134930 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b930af4-f185-44c9-a531-b0313021e3a1","Type":"ContainerDied","Data":"2e9b3c7bbd7dfeab829b80e0e004d8146f9e37ef4f1b06df54a03a1b71b10daa"} Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.135429 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b930af4-f185-44c9-a531-b0313021e3a1","Type":"ContainerDied","Data":"3d791905de8c15ce67c74e8763624bc7e895adc21c973f512c24cbfc8b64aada"} Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.259893 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.321187 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b930af4-f185-44c9-a531-b0313021e3a1-logs\") pod \"4b930af4-f185-44c9-a531-b0313021e3a1\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.321747 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b930af4-f185-44c9-a531-b0313021e3a1-logs" (OuterVolumeSpecName: "logs") pod "4b930af4-f185-44c9-a531-b0313021e3a1" (UID: "4b930af4-f185-44c9-a531-b0313021e3a1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.321961 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj77n\" (UniqueName: \"kubernetes.io/projected/4b930af4-f185-44c9-a531-b0313021e3a1-kube-api-access-nj77n\") pod \"4b930af4-f185-44c9-a531-b0313021e3a1\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.322773 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-combined-ca-bundle\") pod \"4b930af4-f185-44c9-a531-b0313021e3a1\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.322831 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-config-data\") pod \"4b930af4-f185-44c9-a531-b0313021e3a1\" (UID: \"4b930af4-f185-44c9-a531-b0313021e3a1\") " Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.323290 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b930af4-f185-44c9-a531-b0313021e3a1-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.344840 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b930af4-f185-44c9-a531-b0313021e3a1-kube-api-access-nj77n" (OuterVolumeSpecName: "kube-api-access-nj77n") pod "4b930af4-f185-44c9-a531-b0313021e3a1" (UID: "4b930af4-f185-44c9-a531-b0313021e3a1"). InnerVolumeSpecName "kube-api-access-nj77n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.361939 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b930af4-f185-44c9-a531-b0313021e3a1" (UID: "4b930af4-f185-44c9-a531-b0313021e3a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.362110 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-config-data" (OuterVolumeSpecName: "config-data") pod "4b930af4-f185-44c9-a531-b0313021e3a1" (UID: "4b930af4-f185-44c9-a531-b0313021e3a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.424872 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj77n\" (UniqueName: \"kubernetes.io/projected/4b930af4-f185-44c9-a531-b0313021e3a1-kube-api-access-nj77n\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.424922 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:36 crc kubenswrapper[4936]: I0320 16:21:36.424934 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b930af4-f185-44c9-a531-b0313021e3a1-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.151808 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b930af4-f185-44c9-a531-b0313021e3a1","Type":"ContainerDied","Data":"1805a0a9e5a962df02289e8d56249f04ead2a5fee8b0e31a188449fc442360b2"} Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.152282 4936 scope.go:117] "RemoveContainer" containerID="2e9b3c7bbd7dfeab829b80e0e004d8146f9e37ef4f1b06df54a03a1b71b10daa" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.152595 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.177512 4936 scope.go:117] "RemoveContainer" containerID="3d791905de8c15ce67c74e8763624bc7e895adc21c973f512c24cbfc8b64aada" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.200492 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.215274 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.228653 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:37 crc kubenswrapper[4936]: E0320 16:21:37.229148 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b930af4-f185-44c9-a531-b0313021e3a1" containerName="nova-metadata-metadata" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.229178 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b930af4-f185-44c9-a531-b0313021e3a1" containerName="nova-metadata-metadata" Mar 20 16:21:37 crc kubenswrapper[4936]: E0320 16:21:37.229244 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b930af4-f185-44c9-a531-b0313021e3a1" containerName="nova-metadata-log" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.229253 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b930af4-f185-44c9-a531-b0313021e3a1" containerName="nova-metadata-log" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.229496 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b930af4-f185-44c9-a531-b0313021e3a1" containerName="nova-metadata-metadata" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.229524 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b930af4-f185-44c9-a531-b0313021e3a1" containerName="nova-metadata-log" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.230606 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.233139 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.233423 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.239271 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.344102 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3ca600e-664f-4652-924c-106838863904-logs\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.344606 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-config-data\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.344919 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsskx\" (UniqueName: \"kubernetes.io/projected/a3ca600e-664f-4652-924c-106838863904-kube-api-access-hsskx\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.345106 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.345305 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.448168 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.448277 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3ca600e-664f-4652-924c-106838863904-logs\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.448311 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-config-data\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.448475 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsskx\" (UniqueName: \"kubernetes.io/projected/a3ca600e-664f-4652-924c-106838863904-kube-api-access-hsskx\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.448605 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.449534 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3ca600e-664f-4652-924c-106838863904-logs\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.455362 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.456288 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.459492 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-config-data\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.465770 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsskx\" (UniqueName: \"kubernetes.io/projected/a3ca600e-664f-4652-924c-106838863904-kube-api-access-hsskx\") pod \"nova-metadata-0\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.550591 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:37 crc kubenswrapper[4936]: I0320 16:21:37.868691 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b930af4-f185-44c9-a531-b0313021e3a1" path="/var/lib/kubelet/pods/4b930af4-f185-44c9-a531-b0313021e3a1/volumes" Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.074262 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.162232 4936 generic.go:334] "Generic (PLEG): container finished" podID="446d671c-a5e1-4430-9aa6-262bb3a9b2d0" containerID="c18db7a3d9576592f70a01ee0599f74b09c962beacbd3cf95e3dfeeaf1d6b493" exitCode=0 Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.162673 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k54pt" event={"ID":"446d671c-a5e1-4430-9aa6-262bb3a9b2d0","Type":"ContainerDied","Data":"c18db7a3d9576592f70a01ee0599f74b09c962beacbd3cf95e3dfeeaf1d6b493"} Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.167299 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a3ca600e-664f-4652-924c-106838863904","Type":"ContainerStarted","Data":"1ff75bb5c1dc80215ad45169a8abb37d7f689c05fff10477935b7082f7f0942d"} Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.547571 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.596809 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.596851 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.639611 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.797328 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.797447 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.812759 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.871776 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lndxt"] Mar 20 16:21:38 crc kubenswrapper[4936]: I0320 16:21:38.872229 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" podUID="b2a56dab-8614-4f51-bbe7-46f59dda60de" containerName="dnsmasq-dns" containerID="cri-o://4366334e4df96fce7c66687081cbd5d7d49b81f250b372b34a4f3a676ab5f807" gracePeriod=10 Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.190315 4936 generic.go:334] "Generic (PLEG): container finished" podID="b2a56dab-8614-4f51-bbe7-46f59dda60de" containerID="4366334e4df96fce7c66687081cbd5d7d49b81f250b372b34a4f3a676ab5f807" exitCode=0 Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.190365 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" event={"ID":"b2a56dab-8614-4f51-bbe7-46f59dda60de","Type":"ContainerDied","Data":"4366334e4df96fce7c66687081cbd5d7d49b81f250b372b34a4f3a676ab5f807"} Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.199154 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a3ca600e-664f-4652-924c-106838863904","Type":"ContainerStarted","Data":"0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f"} Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.199197 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a3ca600e-664f-4652-924c-106838863904","Type":"ContainerStarted","Data":"4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08"} Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.223985 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.223962326 podStartE2EDuration="2.223962326s" podCreationTimestamp="2026-03-20 16:21:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:39.216097732 +0000 UTC m=+1250.162465557" watchObservedRunningTime="2026-03-20 16:21:39.223962326 +0000 UTC m=+1250.170330151" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.237857 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.706649 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.714920 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798197 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-config-data\") pod \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798323 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-scripts\") pod \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798352 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sr8t\" (UniqueName: \"kubernetes.io/projected/b2a56dab-8614-4f51-bbe7-46f59dda60de-kube-api-access-4sr8t\") pod \"b2a56dab-8614-4f51-bbe7-46f59dda60de\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798383 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-nb\") pod \"b2a56dab-8614-4f51-bbe7-46f59dda60de\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798508 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-swift-storage-0\") pod \"b2a56dab-8614-4f51-bbe7-46f59dda60de\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798577 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-svc\") pod \"b2a56dab-8614-4f51-bbe7-46f59dda60de\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798673 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-sb\") pod \"b2a56dab-8614-4f51-bbe7-46f59dda60de\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798727 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-combined-ca-bundle\") pod \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798792 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8s79\" (UniqueName: \"kubernetes.io/projected/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-kube-api-access-z8s79\") pod \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\" (UID: \"446d671c-a5e1-4430-9aa6-262bb3a9b2d0\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.798861 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-config\") pod \"b2a56dab-8614-4f51-bbe7-46f59dda60de\" (UID: \"b2a56dab-8614-4f51-bbe7-46f59dda60de\") " Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.801073 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.808768 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a56dab-8614-4f51-bbe7-46f59dda60de-kube-api-access-4sr8t" (OuterVolumeSpecName: "kube-api-access-4sr8t") pod "b2a56dab-8614-4f51-bbe7-46f59dda60de" (UID: "b2a56dab-8614-4f51-bbe7-46f59dda60de"). InnerVolumeSpecName "kube-api-access-4sr8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.820396 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-kube-api-access-z8s79" (OuterVolumeSpecName: "kube-api-access-z8s79") pod "446d671c-a5e1-4430-9aa6-262bb3a9b2d0" (UID: "446d671c-a5e1-4430-9aa6-262bb3a9b2d0"). InnerVolumeSpecName "kube-api-access-z8s79". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.837221 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "446d671c-a5e1-4430-9aa6-262bb3a9b2d0" (UID: "446d671c-a5e1-4430-9aa6-262bb3a9b2d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.853798 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.867995 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-scripts" (OuterVolumeSpecName: "scripts") pod "446d671c-a5e1-4430-9aa6-262bb3a9b2d0" (UID: "446d671c-a5e1-4430-9aa6-262bb3a9b2d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.881916 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-config-data" (OuterVolumeSpecName: "config-data") pod "446d671c-a5e1-4430-9aa6-262bb3a9b2d0" (UID: "446d671c-a5e1-4430-9aa6-262bb3a9b2d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.889495 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2a56dab-8614-4f51-bbe7-46f59dda60de" (UID: "b2a56dab-8614-4f51-bbe7-46f59dda60de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.901307 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.901338 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8s79\" (UniqueName: \"kubernetes.io/projected/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-kube-api-access-z8s79\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.901351 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.901367 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446d671c-a5e1-4430-9aa6-262bb3a9b2d0-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.901378 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sr8t\" (UniqueName: \"kubernetes.io/projected/b2a56dab-8614-4f51-bbe7-46f59dda60de-kube-api-access-4sr8t\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.901390 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.905232 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b2a56dab-8614-4f51-bbe7-46f59dda60de" (UID: "b2a56dab-8614-4f51-bbe7-46f59dda60de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.906097 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b2a56dab-8614-4f51-bbe7-46f59dda60de" (UID: "b2a56dab-8614-4f51-bbe7-46f59dda60de"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.913436 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b2a56dab-8614-4f51-bbe7-46f59dda60de" (UID: "b2a56dab-8614-4f51-bbe7-46f59dda60de"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:21:39 crc kubenswrapper[4936]: I0320 16:21:39.926394 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-config" (OuterVolumeSpecName: "config") pod "b2a56dab-8614-4f51-bbe7-46f59dda60de" (UID: "b2a56dab-8614-4f51-bbe7-46f59dda60de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.003070 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.003124 4936 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.003135 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.003146 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a56dab-8614-4f51-bbe7-46f59dda60de-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.212914 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" event={"ID":"b2a56dab-8614-4f51-bbe7-46f59dda60de","Type":"ContainerDied","Data":"ec0a4f6d86ce6d6a243d3bd4a3022259c8c960ffe21ecc6d7702082c7f2669f4"} Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.213016 4936 scope.go:117] "RemoveContainer" containerID="4366334e4df96fce7c66687081cbd5d7d49b81f250b372b34a4f3a676ab5f807" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.214902 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-lndxt" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.214967 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k54pt" event={"ID":"446d671c-a5e1-4430-9aa6-262bb3a9b2d0","Type":"ContainerDied","Data":"749162c9811e511d80bfc53d7b063a0e0cf5110c719edb86dd2d565e8155f139"} Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.214992 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="749162c9811e511d80bfc53d7b063a0e0cf5110c719edb86dd2d565e8155f139" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.215349 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k54pt" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.249797 4936 scope.go:117] "RemoveContainer" containerID="ca32faf9fcc7baa65af0e7c5ef5eb0a4a30b2e8d5d202bc7701f2b08d53b9ee9" Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.275211 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lndxt"] Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.303415 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lndxt"] Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.334648 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.335008 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-log" containerID="cri-o://1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d" gracePeriod=30 Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.335501 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-api" containerID="cri-o://a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf" gracePeriod=30 Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.372082 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:21:40 crc kubenswrapper[4936]: I0320 16:21:40.382444 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:40 crc kubenswrapper[4936]: E0320 16:21:40.515128 4936 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod446d671c_a5e1_4430_9aa6_262bb3a9b2d0.slice\": RecentStats: unable to find data in memory cache]" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.225408 4936 generic.go:334] "Generic (PLEG): container finished" podID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerID="1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d" exitCode=143 Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.225457 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5","Type":"ContainerDied","Data":"1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d"} Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.226968 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="74ae4b9f-ac00-44f8-b072-b65bc3bc4373" containerName="nova-scheduler-scheduler" containerID="cri-o://c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623" gracePeriod=30 Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.227446 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a3ca600e-664f-4652-924c-106838863904" containerName="nova-metadata-log" containerID="cri-o://4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08" gracePeriod=30 Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.227856 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a3ca600e-664f-4652-924c-106838863904" containerName="nova-metadata-metadata" containerID="cri-o://0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f" gracePeriod=30 Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.765717 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.837715 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3ca600e-664f-4652-924c-106838863904-logs\") pod \"a3ca600e-664f-4652-924c-106838863904\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.837915 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-nova-metadata-tls-certs\") pod \"a3ca600e-664f-4652-924c-106838863904\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.837959 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-config-data\") pod \"a3ca600e-664f-4652-924c-106838863904\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.837994 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-combined-ca-bundle\") pod \"a3ca600e-664f-4652-924c-106838863904\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.838094 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsskx\" (UniqueName: \"kubernetes.io/projected/a3ca600e-664f-4652-924c-106838863904-kube-api-access-hsskx\") pod \"a3ca600e-664f-4652-924c-106838863904\" (UID: \"a3ca600e-664f-4652-924c-106838863904\") " Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.838828 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3ca600e-664f-4652-924c-106838863904-logs" (OuterVolumeSpecName: "logs") pod "a3ca600e-664f-4652-924c-106838863904" (UID: "a3ca600e-664f-4652-924c-106838863904"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.843484 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3ca600e-664f-4652-924c-106838863904-kube-api-access-hsskx" (OuterVolumeSpecName: "kube-api-access-hsskx") pod "a3ca600e-664f-4652-924c-106838863904" (UID: "a3ca600e-664f-4652-924c-106838863904"). InnerVolumeSpecName "kube-api-access-hsskx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.867283 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a56dab-8614-4f51-bbe7-46f59dda60de" path="/var/lib/kubelet/pods/b2a56dab-8614-4f51-bbe7-46f59dda60de/volumes" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.875774 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3ca600e-664f-4652-924c-106838863904" (UID: "a3ca600e-664f-4652-924c-106838863904"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.888745 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-config-data" (OuterVolumeSpecName: "config-data") pod "a3ca600e-664f-4652-924c-106838863904" (UID: "a3ca600e-664f-4652-924c-106838863904"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.896892 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a3ca600e-664f-4652-924c-106838863904" (UID: "a3ca600e-664f-4652-924c-106838863904"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.940522 4936 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.940567 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.940578 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ca600e-664f-4652-924c-106838863904-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.940638 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsskx\" (UniqueName: \"kubernetes.io/projected/a3ca600e-664f-4652-924c-106838863904-kube-api-access-hsskx\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:41 crc kubenswrapper[4936]: I0320 16:21:41.940646 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3ca600e-664f-4652-924c-106838863904-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.237240 4936 generic.go:334] "Generic (PLEG): container finished" podID="a3ca600e-664f-4652-924c-106838863904" containerID="0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f" exitCode=0 Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.237534 4936 generic.go:334] "Generic (PLEG): container finished" podID="a3ca600e-664f-4652-924c-106838863904" containerID="4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08" exitCode=143 Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.237346 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a3ca600e-664f-4652-924c-106838863904","Type":"ContainerDied","Data":"0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f"} Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.237591 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a3ca600e-664f-4652-924c-106838863904","Type":"ContainerDied","Data":"4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08"} Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.237610 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a3ca600e-664f-4652-924c-106838863904","Type":"ContainerDied","Data":"1ff75bb5c1dc80215ad45169a8abb37d7f689c05fff10477935b7082f7f0942d"} Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.237633 4936 scope.go:117] "RemoveContainer" containerID="0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.237698 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.263231 4936 scope.go:117] "RemoveContainer" containerID="4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.292452 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.305038 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.316904 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:42 crc kubenswrapper[4936]: E0320 16:21:42.318233 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a56dab-8614-4f51-bbe7-46f59dda60de" containerName="init" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.318339 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a56dab-8614-4f51-bbe7-46f59dda60de" containerName="init" Mar 20 16:21:42 crc kubenswrapper[4936]: E0320 16:21:42.318452 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446d671c-a5e1-4430-9aa6-262bb3a9b2d0" containerName="nova-manage" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.318533 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="446d671c-a5e1-4430-9aa6-262bb3a9b2d0" containerName="nova-manage" Mar 20 16:21:42 crc kubenswrapper[4936]: E0320 16:21:42.318672 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3ca600e-664f-4652-924c-106838863904" containerName="nova-metadata-metadata" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.318748 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3ca600e-664f-4652-924c-106838863904" containerName="nova-metadata-metadata" Mar 20 16:21:42 crc kubenswrapper[4936]: E0320 16:21:42.318880 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a56dab-8614-4f51-bbe7-46f59dda60de" containerName="dnsmasq-dns" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.318957 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a56dab-8614-4f51-bbe7-46f59dda60de" containerName="dnsmasq-dns" Mar 20 16:21:42 crc kubenswrapper[4936]: E0320 16:21:42.319055 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3ca600e-664f-4652-924c-106838863904" containerName="nova-metadata-log" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.319132 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3ca600e-664f-4652-924c-106838863904" containerName="nova-metadata-log" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.319681 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3ca600e-664f-4652-924c-106838863904" containerName="nova-metadata-metadata" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.323771 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3ca600e-664f-4652-924c-106838863904" containerName="nova-metadata-log" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.323824 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a56dab-8614-4f51-bbe7-46f59dda60de" containerName="dnsmasq-dns" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.323861 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="446d671c-a5e1-4430-9aa6-262bb3a9b2d0" containerName="nova-manage" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.329240 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.339290 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.350015 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.356843 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4180e1-8bc4-4b72-9d32-323b920e28e3-logs\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.357080 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-config-data\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.357544 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngjf5\" (UniqueName: \"kubernetes.io/projected/1a4180e1-8bc4-4b72-9d32-323b920e28e3-kube-api-access-ngjf5\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.357795 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.357867 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.368126 4936 scope.go:117] "RemoveContainer" containerID="0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f" Mar 20 16:21:42 crc kubenswrapper[4936]: E0320 16:21:42.368610 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f\": container with ID starting with 0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f not found: ID does not exist" containerID="0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.368653 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f"} err="failed to get container status \"0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f\": rpc error: code = NotFound desc = could not find container \"0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f\": container with ID starting with 0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f not found: ID does not exist" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.368682 4936 scope.go:117] "RemoveContainer" containerID="4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08" Mar 20 16:21:42 crc kubenswrapper[4936]: E0320 16:21:42.369045 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08\": container with ID starting with 4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08 not found: ID does not exist" containerID="4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.369080 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08"} err="failed to get container status \"4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08\": rpc error: code = NotFound desc = could not find container \"4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08\": container with ID starting with 4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08 not found: ID does not exist" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.369122 4936 scope.go:117] "RemoveContainer" containerID="0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.369376 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f"} err="failed to get container status \"0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f\": rpc error: code = NotFound desc = could not find container \"0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f\": container with ID starting with 0695cffe318fc6d7ef72e305c6f1e3147f7f49da27c8425edc621816dc451e6f not found: ID does not exist" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.369397 4936 scope.go:117] "RemoveContainer" containerID="4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.369756 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08"} err="failed to get container status \"4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08\": rpc error: code = NotFound desc = could not find container \"4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08\": container with ID starting with 4a06357e7353aac7fb64f2b85e6e540e060280397b40f64c6cb03012672e1d08 not found: ID does not exist" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.384356 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.459420 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngjf5\" (UniqueName: \"kubernetes.io/projected/1a4180e1-8bc4-4b72-9d32-323b920e28e3-kube-api-access-ngjf5\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.459515 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.459547 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.459658 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4180e1-8bc4-4b72-9d32-323b920e28e3-logs\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.459841 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-config-data\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.460218 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4180e1-8bc4-4b72-9d32-323b920e28e3-logs\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.464714 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.464749 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.465501 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-config-data\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.480437 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngjf5\" (UniqueName: \"kubernetes.io/projected/1a4180e1-8bc4-4b72-9d32-323b920e28e3-kube-api-access-ngjf5\") pod \"nova-metadata-0\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.671844 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.879529 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.970032 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-config-data\") pod \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.970169 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-combined-ca-bundle\") pod \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.970467 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7flt\" (UniqueName: \"kubernetes.io/projected/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-kube-api-access-z7flt\") pod \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\" (UID: \"74ae4b9f-ac00-44f8-b072-b65bc3bc4373\") " Mar 20 16:21:42 crc kubenswrapper[4936]: I0320 16:21:42.983850 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-kube-api-access-z7flt" (OuterVolumeSpecName: "kube-api-access-z7flt") pod "74ae4b9f-ac00-44f8-b072-b65bc3bc4373" (UID: "74ae4b9f-ac00-44f8-b072-b65bc3bc4373"). InnerVolumeSpecName "kube-api-access-z7flt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.000076 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-config-data" (OuterVolumeSpecName: "config-data") pod "74ae4b9f-ac00-44f8-b072-b65bc3bc4373" (UID: "74ae4b9f-ac00-44f8-b072-b65bc3bc4373"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.001473 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74ae4b9f-ac00-44f8-b072-b65bc3bc4373" (UID: "74ae4b9f-ac00-44f8-b072-b65bc3bc4373"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.073289 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7flt\" (UniqueName: \"kubernetes.io/projected/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-kube-api-access-z7flt\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.073338 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.073351 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ae4b9f-ac00-44f8-b072-b65bc3bc4373-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.152509 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.248447 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4180e1-8bc4-4b72-9d32-323b920e28e3","Type":"ContainerStarted","Data":"f6ec3f795c9d42c4988347fce54278a7badfd54a4c45e9463d81fdd697b68761"} Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.251455 4936 generic.go:334] "Generic (PLEG): container finished" podID="74ae4b9f-ac00-44f8-b072-b65bc3bc4373" containerID="c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623" exitCode=0 Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.251511 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"74ae4b9f-ac00-44f8-b072-b65bc3bc4373","Type":"ContainerDied","Data":"c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623"} Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.251544 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"74ae4b9f-ac00-44f8-b072-b65bc3bc4373","Type":"ContainerDied","Data":"1ea311a8a00069b5f9b0f06b2034b5bda4937c281a8ccdde2f793cfa7410cb63"} Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.251612 4936 scope.go:117] "RemoveContainer" containerID="c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.251748 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.287184 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.288316 4936 scope.go:117] "RemoveContainer" containerID="c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623" Mar 20 16:21:43 crc kubenswrapper[4936]: E0320 16:21:43.289143 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623\": container with ID starting with c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623 not found: ID does not exist" containerID="c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.289222 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623"} err="failed to get container status \"c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623\": rpc error: code = NotFound desc = could not find container \"c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623\": container with ID starting with c68df529e2f03c05beb6b9d3630a05aaf8a6b66eb3ba4beab5b3f114c3680623 not found: ID does not exist" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.297314 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.311732 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:21:43 crc kubenswrapper[4936]: E0320 16:21:43.312149 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74ae4b9f-ac00-44f8-b072-b65bc3bc4373" containerName="nova-scheduler-scheduler" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.312173 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="74ae4b9f-ac00-44f8-b072-b65bc3bc4373" containerName="nova-scheduler-scheduler" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.312422 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="74ae4b9f-ac00-44f8-b072-b65bc3bc4373" containerName="nova-scheduler-scheduler" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.313006 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.316122 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.341967 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.379807 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-config-data\") pod \"nova-scheduler-0\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.379975 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.380263 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sq75\" (UniqueName: \"kubernetes.io/projected/265ab68a-88f2-47c2-88d6-9afe238f1733-kube-api-access-6sq75\") pod \"nova-scheduler-0\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.481908 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-config-data\") pod \"nova-scheduler-0\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.481996 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.482119 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sq75\" (UniqueName: \"kubernetes.io/projected/265ab68a-88f2-47c2-88d6-9afe238f1733-kube-api-access-6sq75\") pod \"nova-scheduler-0\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.487115 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-config-data\") pod \"nova-scheduler-0\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.487172 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.501258 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sq75\" (UniqueName: \"kubernetes.io/projected/265ab68a-88f2-47c2-88d6-9afe238f1733-kube-api-access-6sq75\") pod \"nova-scheduler-0\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.632565 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.875016 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74ae4b9f-ac00-44f8-b072-b65bc3bc4373" path="/var/lib/kubelet/pods/74ae4b9f-ac00-44f8-b072-b65bc3bc4373/volumes" Mar 20 16:21:43 crc kubenswrapper[4936]: I0320 16:21:43.875883 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3ca600e-664f-4652-924c-106838863904" path="/var/lib/kubelet/pods/a3ca600e-664f-4652-924c-106838863904/volumes" Mar 20 16:21:44 crc kubenswrapper[4936]: I0320 16:21:44.104450 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:21:44 crc kubenswrapper[4936]: I0320 16:21:44.286921 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4180e1-8bc4-4b72-9d32-323b920e28e3","Type":"ContainerStarted","Data":"8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708"} Mar 20 16:21:44 crc kubenswrapper[4936]: I0320 16:21:44.286971 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4180e1-8bc4-4b72-9d32-323b920e28e3","Type":"ContainerStarted","Data":"84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de"} Mar 20 16:21:44 crc kubenswrapper[4936]: I0320 16:21:44.293070 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"265ab68a-88f2-47c2-88d6-9afe238f1733","Type":"ContainerStarted","Data":"a589bd3309476876c16fb47df8674cddcc00542217e31c3baa0722d9512fb5ba"} Mar 20 16:21:44 crc kubenswrapper[4936]: I0320 16:21:44.312172 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.312152916 podStartE2EDuration="2.312152916s" podCreationTimestamp="2026-03-20 16:21:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:44.309070509 +0000 UTC m=+1255.255438334" watchObservedRunningTime="2026-03-20 16:21:44.312152916 +0000 UTC m=+1255.258520731" Mar 20 16:21:45 crc kubenswrapper[4936]: I0320 16:21:45.182198 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 20 16:21:45 crc kubenswrapper[4936]: I0320 16:21:45.302152 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"265ab68a-88f2-47c2-88d6-9afe238f1733","Type":"ContainerStarted","Data":"67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6"} Mar 20 16:21:45 crc kubenswrapper[4936]: I0320 16:21:45.324772 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.324750019 podStartE2EDuration="2.324750019s" podCreationTimestamp="2026-03-20 16:21:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:45.324685668 +0000 UTC m=+1256.271053483" watchObservedRunningTime="2026-03-20 16:21:45.324750019 +0000 UTC m=+1256.271117834" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.280657 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.314093 4936 generic.go:334] "Generic (PLEG): container finished" podID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerID="a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf" exitCode=0 Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.314155 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5","Type":"ContainerDied","Data":"a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf"} Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.314181 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5","Type":"ContainerDied","Data":"f5b56938f186cf46b2957f247c5180de46a78dd704c3a4ebd78b8ef3f32ff715"} Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.314197 4936 scope.go:117] "RemoveContainer" containerID="a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.314306 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.322183 4936 generic.go:334] "Generic (PLEG): container finished" podID="6071aca2-12af-427f-88a0-5661b147f6d8" containerID="d07e02914919c0f0a57dcf0f6adfb03c41bf1841fe9fa0a3c9bdbc2026c6cecc" exitCode=0 Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.322726 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fvb75" event={"ID":"6071aca2-12af-427f-88a0-5661b147f6d8","Type":"ContainerDied","Data":"d07e02914919c0f0a57dcf0f6adfb03c41bf1841fe9fa0a3c9bdbc2026c6cecc"} Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.344992 4936 scope.go:117] "RemoveContainer" containerID="1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.352477 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-combined-ca-bundle\") pod \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.352737 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-config-data\") pod \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.352800 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-logs\") pod \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.352840 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpldg\" (UniqueName: \"kubernetes.io/projected/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-kube-api-access-fpldg\") pod \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\" (UID: \"a5e62b86-57b1-4ba2-a56a-70a32eda9cc5\") " Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.355050 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-logs" (OuterVolumeSpecName: "logs") pod "a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" (UID: "a5e62b86-57b1-4ba2-a56a-70a32eda9cc5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.369771 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-kube-api-access-fpldg" (OuterVolumeSpecName: "kube-api-access-fpldg") pod "a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" (UID: "a5e62b86-57b1-4ba2-a56a-70a32eda9cc5"). InnerVolumeSpecName "kube-api-access-fpldg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.369881 4936 scope.go:117] "RemoveContainer" containerID="a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf" Mar 20 16:21:46 crc kubenswrapper[4936]: E0320 16:21:46.375976 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf\": container with ID starting with a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf not found: ID does not exist" containerID="a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.376034 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf"} err="failed to get container status \"a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf\": rpc error: code = NotFound desc = could not find container \"a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf\": container with ID starting with a3f1741c2a106d8e1afb58d7f86653400149a41755e92d68473aaef27abdc1cf not found: ID does not exist" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.376062 4936 scope.go:117] "RemoveContainer" containerID="1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d" Mar 20 16:21:46 crc kubenswrapper[4936]: E0320 16:21:46.376609 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d\": container with ID starting with 1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d not found: ID does not exist" containerID="1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.376653 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d"} err="failed to get container status \"1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d\": rpc error: code = NotFound desc = could not find container \"1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d\": container with ID starting with 1d9d798e717d78e42500fe0503e00f02f3a7a5199c1a670b3c4a0137e6098e1d not found: ID does not exist" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.381058 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" (UID: "a5e62b86-57b1-4ba2-a56a-70a32eda9cc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.386569 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-config-data" (OuterVolumeSpecName: "config-data") pod "a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" (UID: "a5e62b86-57b1-4ba2-a56a-70a32eda9cc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.454831 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.454864 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.454877 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.454891 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpldg\" (UniqueName: \"kubernetes.io/projected/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5-kube-api-access-fpldg\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.654212 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.665118 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.681423 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 20 16:21:46 crc kubenswrapper[4936]: E0320 16:21:46.681948 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-api" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.681974 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-api" Mar 20 16:21:46 crc kubenswrapper[4936]: E0320 16:21:46.681988 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-log" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.681996 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-log" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.682272 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-api" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.682300 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" containerName="nova-api-log" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.683472 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.685782 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.689332 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.759913 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/370791fb-0e04-4268-8377-533a0544019e-logs\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.760118 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-config-data\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.760146 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rvhg\" (UniqueName: \"kubernetes.io/projected/370791fb-0e04-4268-8377-533a0544019e-kube-api-access-4rvhg\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.760199 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.861745 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.862171 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/370791fb-0e04-4268-8377-533a0544019e-logs\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.862208 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-config-data\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.862227 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rvhg\" (UniqueName: \"kubernetes.io/projected/370791fb-0e04-4268-8377-533a0544019e-kube-api-access-4rvhg\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.862753 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/370791fb-0e04-4268-8377-533a0544019e-logs\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.865965 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.865970 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-config-data\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:46 crc kubenswrapper[4936]: I0320 16:21:46.879115 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rvhg\" (UniqueName: \"kubernetes.io/projected/370791fb-0e04-4268-8377-533a0544019e-kube-api-access-4rvhg\") pod \"nova-api-0\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " pod="openstack/nova-api-0" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.009038 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.474079 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.750886 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.873444 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5e62b86-57b1-4ba2-a56a-70a32eda9cc5" path="/var/lib/kubelet/pods/a5e62b86-57b1-4ba2-a56a-70a32eda9cc5/volumes" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.891062 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jf8x\" (UniqueName: \"kubernetes.io/projected/6071aca2-12af-427f-88a0-5661b147f6d8-kube-api-access-7jf8x\") pod \"6071aca2-12af-427f-88a0-5661b147f6d8\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.891471 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-scripts\") pod \"6071aca2-12af-427f-88a0-5661b147f6d8\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.891534 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-config-data\") pod \"6071aca2-12af-427f-88a0-5661b147f6d8\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.891682 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-combined-ca-bundle\") pod \"6071aca2-12af-427f-88a0-5661b147f6d8\" (UID: \"6071aca2-12af-427f-88a0-5661b147f6d8\") " Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.895461 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6071aca2-12af-427f-88a0-5661b147f6d8-kube-api-access-7jf8x" (OuterVolumeSpecName: "kube-api-access-7jf8x") pod "6071aca2-12af-427f-88a0-5661b147f6d8" (UID: "6071aca2-12af-427f-88a0-5661b147f6d8"). InnerVolumeSpecName "kube-api-access-7jf8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.895754 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-scripts" (OuterVolumeSpecName: "scripts") pod "6071aca2-12af-427f-88a0-5661b147f6d8" (UID: "6071aca2-12af-427f-88a0-5661b147f6d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.924240 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-config-data" (OuterVolumeSpecName: "config-data") pod "6071aca2-12af-427f-88a0-5661b147f6d8" (UID: "6071aca2-12af-427f-88a0-5661b147f6d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.934536 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6071aca2-12af-427f-88a0-5661b147f6d8" (UID: "6071aca2-12af-427f-88a0-5661b147f6d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.994604 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.994653 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.994670 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6071aca2-12af-427f-88a0-5661b147f6d8-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:47 crc kubenswrapper[4936]: I0320 16:21:47.994684 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jf8x\" (UniqueName: \"kubernetes.io/projected/6071aca2-12af-427f-88a0-5661b147f6d8-kube-api-access-7jf8x\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.346487 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fvb75" event={"ID":"6071aca2-12af-427f-88a0-5661b147f6d8","Type":"ContainerDied","Data":"66acb92317c2e5bc49c1cfdc4c7efed8a75d92e13799dd7930dc04127f4e3089"} Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.346880 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66acb92317c2e5bc49c1cfdc4c7efed8a75d92e13799dd7930dc04127f4e3089" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.346567 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fvb75" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.354641 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"370791fb-0e04-4268-8377-533a0544019e","Type":"ContainerStarted","Data":"04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875"} Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.354740 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"370791fb-0e04-4268-8377-533a0544019e","Type":"ContainerStarted","Data":"6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839"} Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.354763 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"370791fb-0e04-4268-8377-533a0544019e","Type":"ContainerStarted","Data":"686c28a3ff11422dd0e209d5fb261d5e27cd844ce992c73f85a1dc5a1b5a3a26"} Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.443220 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.44319823 podStartE2EDuration="2.44319823s" podCreationTimestamp="2026-03-20 16:21:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:48.386595901 +0000 UTC m=+1259.332963736" watchObservedRunningTime="2026-03-20 16:21:48.44319823 +0000 UTC m=+1259.389566045" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.457517 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 20 16:21:48 crc kubenswrapper[4936]: E0320 16:21:48.458116 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6071aca2-12af-427f-88a0-5661b147f6d8" containerName="nova-cell1-conductor-db-sync" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.458139 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="6071aca2-12af-427f-88a0-5661b147f6d8" containerName="nova-cell1-conductor-db-sync" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.458311 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="6071aca2-12af-427f-88a0-5661b147f6d8" containerName="nova-cell1-conductor-db-sync" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.459112 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.461979 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.467356 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.510453 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7810d8f5-e6ff-402c-a916-42e6125f2d6e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7810d8f5-e6ff-402c-a916-42e6125f2d6e\") " pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.510607 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7810d8f5-e6ff-402c-a916-42e6125f2d6e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7810d8f5-e6ff-402c-a916-42e6125f2d6e\") " pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.510635 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgds8\" (UniqueName: \"kubernetes.io/projected/7810d8f5-e6ff-402c-a916-42e6125f2d6e-kube-api-access-jgds8\") pod \"nova-cell1-conductor-0\" (UID: \"7810d8f5-e6ff-402c-a916-42e6125f2d6e\") " pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.612834 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7810d8f5-e6ff-402c-a916-42e6125f2d6e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7810d8f5-e6ff-402c-a916-42e6125f2d6e\") " pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.613012 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7810d8f5-e6ff-402c-a916-42e6125f2d6e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7810d8f5-e6ff-402c-a916-42e6125f2d6e\") " pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.613041 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgds8\" (UniqueName: \"kubernetes.io/projected/7810d8f5-e6ff-402c-a916-42e6125f2d6e-kube-api-access-jgds8\") pod \"nova-cell1-conductor-0\" (UID: \"7810d8f5-e6ff-402c-a916-42e6125f2d6e\") " pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.619151 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7810d8f5-e6ff-402c-a916-42e6125f2d6e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7810d8f5-e6ff-402c-a916-42e6125f2d6e\") " pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.619291 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7810d8f5-e6ff-402c-a916-42e6125f2d6e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7810d8f5-e6ff-402c-a916-42e6125f2d6e\") " pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.629597 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgds8\" (UniqueName: \"kubernetes.io/projected/7810d8f5-e6ff-402c-a916-42e6125f2d6e-kube-api-access-jgds8\") pod \"nova-cell1-conductor-0\" (UID: \"7810d8f5-e6ff-402c-a916-42e6125f2d6e\") " pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.633018 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 20 16:21:48 crc kubenswrapper[4936]: I0320 16:21:48.779156 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.129752 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.130229 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="eff64cd4-dfe6-438f-aed8-4c1637cc2db4" containerName="kube-state-metrics" containerID="cri-o://dd91ee26f881be46e6bcd680946f0538c4007735d7ffb3f683b2b971e71ed118" gracePeriod=30 Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.204918 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 20 16:21:49 crc kubenswrapper[4936]: W0320 16:21:49.218245 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7810d8f5_e6ff_402c_a916_42e6125f2d6e.slice/crio-f41169b7959e67e9ea7ec61613457a6c8e3109cdef86f0564e9b9dc055d2a4e1 WatchSource:0}: Error finding container f41169b7959e67e9ea7ec61613457a6c8e3109cdef86f0564e9b9dc055d2a4e1: Status 404 returned error can't find the container with id f41169b7959e67e9ea7ec61613457a6c8e3109cdef86f0564e9b9dc055d2a4e1 Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.367499 4936 generic.go:334] "Generic (PLEG): container finished" podID="eff64cd4-dfe6-438f-aed8-4c1637cc2db4" containerID="dd91ee26f881be46e6bcd680946f0538c4007735d7ffb3f683b2b971e71ed118" exitCode=2 Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.367582 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eff64cd4-dfe6-438f-aed8-4c1637cc2db4","Type":"ContainerDied","Data":"dd91ee26f881be46e6bcd680946f0538c4007735d7ffb3f683b2b971e71ed118"} Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.368930 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7810d8f5-e6ff-402c-a916-42e6125f2d6e","Type":"ContainerStarted","Data":"f41169b7959e67e9ea7ec61613457a6c8e3109cdef86f0564e9b9dc055d2a4e1"} Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.609511 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.732356 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-747tw\" (UniqueName: \"kubernetes.io/projected/eff64cd4-dfe6-438f-aed8-4c1637cc2db4-kube-api-access-747tw\") pod \"eff64cd4-dfe6-438f-aed8-4c1637cc2db4\" (UID: \"eff64cd4-dfe6-438f-aed8-4c1637cc2db4\") " Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.737113 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff64cd4-dfe6-438f-aed8-4c1637cc2db4-kube-api-access-747tw" (OuterVolumeSpecName: "kube-api-access-747tw") pod "eff64cd4-dfe6-438f-aed8-4c1637cc2db4" (UID: "eff64cd4-dfe6-438f-aed8-4c1637cc2db4"). InnerVolumeSpecName "kube-api-access-747tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:49 crc kubenswrapper[4936]: I0320 16:21:49.835991 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-747tw\" (UniqueName: \"kubernetes.io/projected/eff64cd4-dfe6-438f-aed8-4c1637cc2db4-kube-api-access-747tw\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.384718 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eff64cd4-dfe6-438f-aed8-4c1637cc2db4","Type":"ContainerDied","Data":"74af9ce3a097763ae1a49769cd4660d866d6f4afbe4834cafcdb215e8a4f02f6"} Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.384801 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.384844 4936 scope.go:117] "RemoveContainer" containerID="dd91ee26f881be46e6bcd680946f0538c4007735d7ffb3f683b2b971e71ed118" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.391297 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7810d8f5-e6ff-402c-a916-42e6125f2d6e","Type":"ContainerStarted","Data":"58e899489336f6f70ff9b0c6a172d5488f3ac3a12e69575d5905a3a494fe7be7"} Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.392625 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.410880 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.421850 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.450723 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.450700233 podStartE2EDuration="2.450700233s" podCreationTimestamp="2026-03-20 16:21:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:21:50.438164666 +0000 UTC m=+1261.384532481" watchObservedRunningTime="2026-03-20 16:21:50.450700233 +0000 UTC m=+1261.397068048" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.470569 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 20 16:21:50 crc kubenswrapper[4936]: E0320 16:21:50.471252 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff64cd4-dfe6-438f-aed8-4c1637cc2db4" containerName="kube-state-metrics" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.471387 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff64cd4-dfe6-438f-aed8-4c1637cc2db4" containerName="kube-state-metrics" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.471784 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff64cd4-dfe6-438f-aed8-4c1637cc2db4" containerName="kube-state-metrics" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.472636 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.475635 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.475783 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.486330 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.548828 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc720ee-42c3-4ce6-87de-045d881d730d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.548968 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4cc720ee-42c3-4ce6-87de-045d881d730d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.549007 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4cp2\" (UniqueName: \"kubernetes.io/projected/4cc720ee-42c3-4ce6-87de-045d881d730d-kube-api-access-k4cp2\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.549029 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc720ee-42c3-4ce6-87de-045d881d730d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.651874 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4cc720ee-42c3-4ce6-87de-045d881d730d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.651956 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4cp2\" (UniqueName: \"kubernetes.io/projected/4cc720ee-42c3-4ce6-87de-045d881d730d-kube-api-access-k4cp2\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.651992 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc720ee-42c3-4ce6-87de-045d881d730d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.652067 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc720ee-42c3-4ce6-87de-045d881d730d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.657668 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc720ee-42c3-4ce6-87de-045d881d730d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.658031 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4cc720ee-42c3-4ce6-87de-045d881d730d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.667394 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc720ee-42c3-4ce6-87de-045d881d730d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.669031 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4cp2\" (UniqueName: \"kubernetes.io/projected/4cc720ee-42c3-4ce6-87de-045d881d730d-kube-api-access-k4cp2\") pod \"kube-state-metrics-0\" (UID: \"4cc720ee-42c3-4ce6-87de-045d881d730d\") " pod="openstack/kube-state-metrics-0" Mar 20 16:21:50 crc kubenswrapper[4936]: I0320 16:21:50.795356 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.098388 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.099020 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="ceilometer-central-agent" containerID="cri-o://5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b" gracePeriod=30 Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.099103 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="sg-core" containerID="cri-o://eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031" gracePeriod=30 Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.099163 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="ceilometer-notification-agent" containerID="cri-o://59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767" gracePeriod=30 Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.099234 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="proxy-httpd" containerID="cri-o://6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4" gracePeriod=30 Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.306932 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.402158 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4cc720ee-42c3-4ce6-87de-045d881d730d","Type":"ContainerStarted","Data":"e4f10de5cd1cb0e5cac9cd77ea873cfd6d9e35e2b317cfe86f2edb0ae0e01b5f"} Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.405400 4936 generic.go:334] "Generic (PLEG): container finished" podID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerID="6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4" exitCode=0 Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.405440 4936 generic.go:334] "Generic (PLEG): container finished" podID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerID="eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031" exitCode=2 Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.405483 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerDied","Data":"6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4"} Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.405527 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerDied","Data":"eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031"} Mar 20 16:21:51 crc kubenswrapper[4936]: I0320 16:21:51.866449 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff64cd4-dfe6-438f-aed8-4c1637cc2db4" path="/var/lib/kubelet/pods/eff64cd4-dfe6-438f-aed8-4c1637cc2db4/volumes" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.300972 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.379795 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-config-data\") pod \"cf000d82-6986-4c78-80bf-7913eeed68b9\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.379853 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-combined-ca-bundle\") pod \"cf000d82-6986-4c78-80bf-7913eeed68b9\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.379905 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-run-httpd\") pod \"cf000d82-6986-4c78-80bf-7913eeed68b9\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.379941 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qdg7\" (UniqueName: \"kubernetes.io/projected/cf000d82-6986-4c78-80bf-7913eeed68b9-kube-api-access-6qdg7\") pod \"cf000d82-6986-4c78-80bf-7913eeed68b9\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.380066 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-scripts\") pod \"cf000d82-6986-4c78-80bf-7913eeed68b9\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.380089 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-log-httpd\") pod \"cf000d82-6986-4c78-80bf-7913eeed68b9\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.380125 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-sg-core-conf-yaml\") pod \"cf000d82-6986-4c78-80bf-7913eeed68b9\" (UID: \"cf000d82-6986-4c78-80bf-7913eeed68b9\") " Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.380685 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cf000d82-6986-4c78-80bf-7913eeed68b9" (UID: "cf000d82-6986-4c78-80bf-7913eeed68b9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.380901 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cf000d82-6986-4c78-80bf-7913eeed68b9" (UID: "cf000d82-6986-4c78-80bf-7913eeed68b9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.384115 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf000d82-6986-4c78-80bf-7913eeed68b9-kube-api-access-6qdg7" (OuterVolumeSpecName: "kube-api-access-6qdg7") pod "cf000d82-6986-4c78-80bf-7913eeed68b9" (UID: "cf000d82-6986-4c78-80bf-7913eeed68b9"). InnerVolumeSpecName "kube-api-access-6qdg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.397215 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-scripts" (OuterVolumeSpecName: "scripts") pod "cf000d82-6986-4c78-80bf-7913eeed68b9" (UID: "cf000d82-6986-4c78-80bf-7913eeed68b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.419978 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4cc720ee-42c3-4ce6-87de-045d881d730d","Type":"ContainerStarted","Data":"5239c45193e85ede4ae23ed526ad08a44111a1f1993e36503d1015140ab5ec67"} Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.421076 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.435570 4936 generic.go:334] "Generic (PLEG): container finished" podID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerID="59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767" exitCode=0 Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.435607 4936 generic.go:334] "Generic (PLEG): container finished" podID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerID="5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b" exitCode=0 Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.435729 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.436187 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerDied","Data":"59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767"} Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.436600 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerDied","Data":"5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b"} Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.436615 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf000d82-6986-4c78-80bf-7913eeed68b9","Type":"ContainerDied","Data":"5d3975c8ea69b0efa6886971825c8aa23fede17c8c0fff103c1e8b534dc95ca2"} Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.436659 4936 scope.go:117] "RemoveContainer" containerID="6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.448736 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.930251669 podStartE2EDuration="2.448704966s" podCreationTimestamp="2026-03-20 16:21:50 +0000 UTC" firstStartedPulling="2026-03-20 16:21:51.3135825 +0000 UTC m=+1262.259950315" lastFinishedPulling="2026-03-20 16:21:51.832035797 +0000 UTC m=+1262.778403612" observedRunningTime="2026-03-20 16:21:52.439741791 +0000 UTC m=+1263.386109626" watchObservedRunningTime="2026-03-20 16:21:52.448704966 +0000 UTC m=+1263.395072781" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.449201 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cf000d82-6986-4c78-80bf-7913eeed68b9" (UID: "cf000d82-6986-4c78-80bf-7913eeed68b9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.483144 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf000d82-6986-4c78-80bf-7913eeed68b9" (UID: "cf000d82-6986-4c78-80bf-7913eeed68b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.483469 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.483504 4936 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.483568 4936 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.483579 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.483588 4936 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf000d82-6986-4c78-80bf-7913eeed68b9-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.483599 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qdg7\" (UniqueName: \"kubernetes.io/projected/cf000d82-6986-4c78-80bf-7913eeed68b9-kube-api-access-6qdg7\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.527602 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-config-data" (OuterVolumeSpecName: "config-data") pod "cf000d82-6986-4c78-80bf-7913eeed68b9" (UID: "cf000d82-6986-4c78-80bf-7913eeed68b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.553175 4936 scope.go:117] "RemoveContainer" containerID="eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.575269 4936 scope.go:117] "RemoveContainer" containerID="59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.586078 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf000d82-6986-4c78-80bf-7913eeed68b9-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.601788 4936 scope.go:117] "RemoveContainer" containerID="5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.624641 4936 scope.go:117] "RemoveContainer" containerID="6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4" Mar 20 16:21:52 crc kubenswrapper[4936]: E0320 16:21:52.625067 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4\": container with ID starting with 6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4 not found: ID does not exist" containerID="6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.625120 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4"} err="failed to get container status \"6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4\": rpc error: code = NotFound desc = could not find container \"6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4\": container with ID starting with 6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4 not found: ID does not exist" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.625150 4936 scope.go:117] "RemoveContainer" containerID="eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031" Mar 20 16:21:52 crc kubenswrapper[4936]: E0320 16:21:52.625450 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031\": container with ID starting with eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031 not found: ID does not exist" containerID="eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.625474 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031"} err="failed to get container status \"eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031\": rpc error: code = NotFound desc = could not find container \"eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031\": container with ID starting with eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031 not found: ID does not exist" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.625487 4936 scope.go:117] "RemoveContainer" containerID="59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767" Mar 20 16:21:52 crc kubenswrapper[4936]: E0320 16:21:52.625716 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767\": container with ID starting with 59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767 not found: ID does not exist" containerID="59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.625752 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767"} err="failed to get container status \"59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767\": rpc error: code = NotFound desc = could not find container \"59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767\": container with ID starting with 59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767 not found: ID does not exist" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.625772 4936 scope.go:117] "RemoveContainer" containerID="5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b" Mar 20 16:21:52 crc kubenswrapper[4936]: E0320 16:21:52.626041 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b\": container with ID starting with 5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b not found: ID does not exist" containerID="5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.626089 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b"} err="failed to get container status \"5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b\": rpc error: code = NotFound desc = could not find container \"5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b\": container with ID starting with 5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b not found: ID does not exist" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.626120 4936 scope.go:117] "RemoveContainer" containerID="6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.626373 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4"} err="failed to get container status \"6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4\": rpc error: code = NotFound desc = could not find container \"6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4\": container with ID starting with 6d97d852024f7aeeb8755d5c5862a3f167336079acae9ed6599dcb186ef895e4 not found: ID does not exist" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.626396 4936 scope.go:117] "RemoveContainer" containerID="eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.626600 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031"} err="failed to get container status \"eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031\": rpc error: code = NotFound desc = could not find container \"eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031\": container with ID starting with eb6199672f566ab7229a493b2600305885413cc4120cc2dda9e43587eb94a031 not found: ID does not exist" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.626619 4936 scope.go:117] "RemoveContainer" containerID="59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.626786 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767"} err="failed to get container status \"59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767\": rpc error: code = NotFound desc = could not find container \"59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767\": container with ID starting with 59a7a61c1e3a2988ff9ee7cf3c9abff4b6f2109029caa61e36b4f4ecd8755767 not found: ID does not exist" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.626803 4936 scope.go:117] "RemoveContainer" containerID="5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.626962 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b"} err="failed to get container status \"5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b\": rpc error: code = NotFound desc = could not find container \"5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b\": container with ID starting with 5a20c6c4ef4ce9cac2830ba62fce8bcf94ae9c8e51685be44e136c22314d530b not found: ID does not exist" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.673807 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.673871 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.782579 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.790924 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.808558 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:52 crc kubenswrapper[4936]: E0320 16:21:52.809014 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="proxy-httpd" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.809039 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="proxy-httpd" Mar 20 16:21:52 crc kubenswrapper[4936]: E0320 16:21:52.809060 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="ceilometer-notification-agent" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.809069 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="ceilometer-notification-agent" Mar 20 16:21:52 crc kubenswrapper[4936]: E0320 16:21:52.809088 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="sg-core" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.809097 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="sg-core" Mar 20 16:21:52 crc kubenswrapper[4936]: E0320 16:21:52.809119 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="ceilometer-central-agent" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.809127 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="ceilometer-central-agent" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.809340 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="sg-core" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.809362 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="proxy-httpd" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.809386 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="ceilometer-notification-agent" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.809399 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" containerName="ceilometer-central-agent" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.811459 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.814793 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.815169 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.815298 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.825282 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.896871 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.898421 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.898579 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-config-data\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.898761 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdfrn\" (UniqueName: \"kubernetes.io/projected/f950892c-8013-41c7-97f2-ca160b1c00dd-kube-api-access-fdfrn\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.898967 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-scripts\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.899153 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.899461 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-log-httpd\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:52 crc kubenswrapper[4936]: I0320 16:21:52.899648 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-run-httpd\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.001639 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.001705 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-log-httpd\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.001743 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-run-httpd\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.001782 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.001832 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.001856 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-config-data\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.001967 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdfrn\" (UniqueName: \"kubernetes.io/projected/f950892c-8013-41c7-97f2-ca160b1c00dd-kube-api-access-fdfrn\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.002022 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-scripts\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.003332 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-run-httpd\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.003368 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-log-httpd\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.006471 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.006532 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.006585 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-scripts\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.007210 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-config-data\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.010820 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.024857 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdfrn\" (UniqueName: \"kubernetes.io/projected/f950892c-8013-41c7-97f2-ca160b1c00dd-kube-api-access-fdfrn\") pod \"ceilometer-0\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.130800 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.553308 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.632947 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.662127 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.683825 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.683860 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 20 16:21:53 crc kubenswrapper[4936]: I0320 16:21:53.874484 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf000d82-6986-4c78-80bf-7913eeed68b9" path="/var/lib/kubelet/pods/cf000d82-6986-4c78-80bf-7913eeed68b9/volumes" Mar 20 16:21:54 crc kubenswrapper[4936]: I0320 16:21:54.463765 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerStarted","Data":"93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37"} Mar 20 16:21:54 crc kubenswrapper[4936]: I0320 16:21:54.463809 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerStarted","Data":"8f91044be16fa217ee587605c3883ce1a268e8f09245e046e679c5953d6f0da2"} Mar 20 16:21:54 crc kubenswrapper[4936]: I0320 16:21:54.495245 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 20 16:21:55 crc kubenswrapper[4936]: I0320 16:21:55.476842 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerStarted","Data":"1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e"} Mar 20 16:21:56 crc kubenswrapper[4936]: I0320 16:21:56.487037 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerStarted","Data":"58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347"} Mar 20 16:21:57 crc kubenswrapper[4936]: I0320 16:21:57.009277 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 20 16:21:57 crc kubenswrapper[4936]: I0320 16:21:57.009340 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 20 16:21:58 crc kubenswrapper[4936]: I0320 16:21:58.091756 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 20 16:21:58 crc kubenswrapper[4936]: I0320 16:21:58.091908 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 20 16:21:58 crc kubenswrapper[4936]: I0320 16:21:58.679250 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerStarted","Data":"170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50"} Mar 20 16:21:58 crc kubenswrapper[4936]: I0320 16:21:58.679651 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 20 16:21:58 crc kubenswrapper[4936]: I0320 16:21:58.722249 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.45550642 podStartE2EDuration="6.722212039s" podCreationTimestamp="2026-03-20 16:21:52 +0000 UTC" firstStartedPulling="2026-03-20 16:21:53.567433356 +0000 UTC m=+1264.513801171" lastFinishedPulling="2026-03-20 16:21:57.834138975 +0000 UTC m=+1268.780506790" observedRunningTime="2026-03-20 16:21:58.714911721 +0000 UTC m=+1269.661279546" watchObservedRunningTime="2026-03-20 16:21:58.722212039 +0000 UTC m=+1269.668579854" Mar 20 16:21:58 crc kubenswrapper[4936]: I0320 16:21:58.810754 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.133044 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567062-b6frx"] Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.134399 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567062-b6frx" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.139351 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.139826 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.140298 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.149625 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567062-b6frx"] Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.198573 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m449\" (UniqueName: \"kubernetes.io/projected/643cf02c-71a1-498a-924d-ff539faf7bc7-kube-api-access-7m449\") pod \"auto-csr-approver-29567062-b6frx\" (UID: \"643cf02c-71a1-498a-924d-ff539faf7bc7\") " pod="openshift-infra/auto-csr-approver-29567062-b6frx" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.300598 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m449\" (UniqueName: \"kubernetes.io/projected/643cf02c-71a1-498a-924d-ff539faf7bc7-kube-api-access-7m449\") pod \"auto-csr-approver-29567062-b6frx\" (UID: \"643cf02c-71a1-498a-924d-ff539faf7bc7\") " pod="openshift-infra/auto-csr-approver-29567062-b6frx" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.339318 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m449\" (UniqueName: \"kubernetes.io/projected/643cf02c-71a1-498a-924d-ff539faf7bc7-kube-api-access-7m449\") pod \"auto-csr-approver-29567062-b6frx\" (UID: \"643cf02c-71a1-498a-924d-ff539faf7bc7\") " pod="openshift-infra/auto-csr-approver-29567062-b6frx" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.460492 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567062-b6frx" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.673453 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.673772 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.805125 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 20 16:22:00 crc kubenswrapper[4936]: I0320 16:22:00.929962 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567062-b6frx"] Mar 20 16:22:01 crc kubenswrapper[4936]: I0320 16:22:01.723879 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567062-b6frx" event={"ID":"643cf02c-71a1-498a-924d-ff539faf7bc7","Type":"ContainerStarted","Data":"d6f265a4dd199417a606656761f1f47dd17b32cb57740c94d32cae4a452d859d"} Mar 20 16:22:02 crc kubenswrapper[4936]: I0320 16:22:02.678172 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 20 16:22:02 crc kubenswrapper[4936]: I0320 16:22:02.680125 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 20 16:22:02 crc kubenswrapper[4936]: I0320 16:22:02.684169 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 20 16:22:02 crc kubenswrapper[4936]: I0320 16:22:02.734429 4936 generic.go:334] "Generic (PLEG): container finished" podID="643cf02c-71a1-498a-924d-ff539faf7bc7" containerID="3d2fce81bfbcdbdd93bcb9cfb855367cdb97743b31bb2d69a5ea82518876816f" exitCode=0 Mar 20 16:22:02 crc kubenswrapper[4936]: I0320 16:22:02.734498 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567062-b6frx" event={"ID":"643cf02c-71a1-498a-924d-ff539faf7bc7","Type":"ContainerDied","Data":"3d2fce81bfbcdbdd93bcb9cfb855367cdb97743b31bb2d69a5ea82518876816f"} Mar 20 16:22:02 crc kubenswrapper[4936]: I0320 16:22:02.742975 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 20 16:22:04 crc kubenswrapper[4936]: I0320 16:22:04.076179 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567062-b6frx" Mar 20 16:22:04 crc kubenswrapper[4936]: I0320 16:22:04.180306 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m449\" (UniqueName: \"kubernetes.io/projected/643cf02c-71a1-498a-924d-ff539faf7bc7-kube-api-access-7m449\") pod \"643cf02c-71a1-498a-924d-ff539faf7bc7\" (UID: \"643cf02c-71a1-498a-924d-ff539faf7bc7\") " Mar 20 16:22:04 crc kubenswrapper[4936]: I0320 16:22:04.196966 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/643cf02c-71a1-498a-924d-ff539faf7bc7-kube-api-access-7m449" (OuterVolumeSpecName: "kube-api-access-7m449") pod "643cf02c-71a1-498a-924d-ff539faf7bc7" (UID: "643cf02c-71a1-498a-924d-ff539faf7bc7"). InnerVolumeSpecName "kube-api-access-7m449". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:04 crc kubenswrapper[4936]: I0320 16:22:04.198450 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m449\" (UniqueName: \"kubernetes.io/projected/643cf02c-71a1-498a-924d-ff539faf7bc7-kube-api-access-7m449\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:04 crc kubenswrapper[4936]: I0320 16:22:04.756017 4936 generic.go:334] "Generic (PLEG): container finished" podID="99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd" containerID="bb10f7f7dbf8aa243ed767f0471e602deeb4ca721880d72deee1eab1e3da0d82" exitCode=137 Mar 20 16:22:04 crc kubenswrapper[4936]: I0320 16:22:04.756479 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd","Type":"ContainerDied","Data":"bb10f7f7dbf8aa243ed767f0471e602deeb4ca721880d72deee1eab1e3da0d82"} Mar 20 16:22:04 crc kubenswrapper[4936]: I0320 16:22:04.767410 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567062-b6frx" Mar 20 16:22:04 crc kubenswrapper[4936]: I0320 16:22:04.767537 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567062-b6frx" event={"ID":"643cf02c-71a1-498a-924d-ff539faf7bc7","Type":"ContainerDied","Data":"d6f265a4dd199417a606656761f1f47dd17b32cb57740c94d32cae4a452d859d"} Mar 20 16:22:04 crc kubenswrapper[4936]: I0320 16:22:04.767610 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6f265a4dd199417a606656761f1f47dd17b32cb57740c94d32cae4a452d859d" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.010069 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.010354 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.038803 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.114577 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-config-data\") pod \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.114811 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-combined-ca-bundle\") pod \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.114966 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjdq7\" (UniqueName: \"kubernetes.io/projected/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-kube-api-access-qjdq7\") pod \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\" (UID: \"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd\") " Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.141115 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-kube-api-access-qjdq7" (OuterVolumeSpecName: "kube-api-access-qjdq7") pod "99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd" (UID: "99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd"). InnerVolumeSpecName "kube-api-access-qjdq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.143599 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-config-data" (OuterVolumeSpecName: "config-data") pod "99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd" (UID: "99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.148668 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567056-n6qb7"] Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.159789 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567056-n6qb7"] Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.163174 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd" (UID: "99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.221284 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.221317 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjdq7\" (UniqueName: \"kubernetes.io/projected/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-kube-api-access-qjdq7\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.221329 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.777766 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd","Type":"ContainerDied","Data":"4afd3d55b76e422b851f7dca01608c7462a4a35f844417f38c6a33f0094f22ef"} Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.777823 4936 scope.go:117] "RemoveContainer" containerID="bb10f7f7dbf8aa243ed767f0471e602deeb4ca721880d72deee1eab1e3da0d82" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.777826 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.815570 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.829415 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.837362 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 20 16:22:05 crc kubenswrapper[4936]: E0320 16:22:05.837738 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd" containerName="nova-cell1-novncproxy-novncproxy" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.837751 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd" containerName="nova-cell1-novncproxy-novncproxy" Mar 20 16:22:05 crc kubenswrapper[4936]: E0320 16:22:05.837768 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643cf02c-71a1-498a-924d-ff539faf7bc7" containerName="oc" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.837775 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="643cf02c-71a1-498a-924d-ff539faf7bc7" containerName="oc" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.837946 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="643cf02c-71a1-498a-924d-ff539faf7bc7" containerName="oc" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.837968 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd" containerName="nova-cell1-novncproxy-novncproxy" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.838516 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.841287 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.841532 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.841733 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.866790 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd" path="/var/lib/kubelet/pods/99ba5b5a-c9c2-4ec0-b6e8-92802dfd1fcd/volumes" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.867303 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6" path="/var/lib/kubelet/pods/bbe1f0c7-6d84-4883-9ce3-c2d9498c80e6/volumes" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.868118 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.932955 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.933033 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.933195 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.933212 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:05 crc kubenswrapper[4936]: I0320 16:22:05.933334 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg258\" (UniqueName: \"kubernetes.io/projected/556eac71-c652-4652-8b3e-2b208ca34704-kube-api-access-zg258\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.035295 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.035347 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.035393 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg258\" (UniqueName: \"kubernetes.io/projected/556eac71-c652-4652-8b3e-2b208ca34704-kube-api-access-zg258\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.035449 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.035498 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.039074 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.039536 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.039812 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.040120 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/556eac71-c652-4652-8b3e-2b208ca34704-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.057824 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg258\" (UniqueName: \"kubernetes.io/projected/556eac71-c652-4652-8b3e-2b208ca34704-kube-api-access-zg258\") pod \"nova-cell1-novncproxy-0\" (UID: \"556eac71-c652-4652-8b3e-2b208ca34704\") " pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.163645 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.619121 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 20 16:22:06 crc kubenswrapper[4936]: W0320 16:22:06.621687 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod556eac71_c652_4652_8b3e_2b208ca34704.slice/crio-6d6436a40a41c08ccae52babbcf7d1a8673e3c13ddc11ce46813dce2b9ed608f WatchSource:0}: Error finding container 6d6436a40a41c08ccae52babbcf7d1a8673e3c13ddc11ce46813dce2b9ed608f: Status 404 returned error can't find the container with id 6d6436a40a41c08ccae52babbcf7d1a8673e3c13ddc11ce46813dce2b9ed608f Mar 20 16:22:06 crc kubenswrapper[4936]: I0320 16:22:06.806406 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"556eac71-c652-4652-8b3e-2b208ca34704","Type":"ContainerStarted","Data":"6d6436a40a41c08ccae52babbcf7d1a8673e3c13ddc11ce46813dce2b9ed608f"} Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.013653 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.014015 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.018315 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.021921 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.300117 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-ln75g"] Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.301516 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.344490 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-ln75g"] Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.382173 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.382223 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.382264 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.382314 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7pb2\" (UniqueName: \"kubernetes.io/projected/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-kube-api-access-z7pb2\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.382352 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.382426 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-config\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.492138 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.492179 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.492222 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.492259 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7pb2\" (UniqueName: \"kubernetes.io/projected/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-kube-api-access-z7pb2\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.492288 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.492339 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-config\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.492939 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.493105 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.495172 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-config\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.495241 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.497676 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.525424 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7pb2\" (UniqueName: \"kubernetes.io/projected/3eed9143-556a-4df3-bef4-1eb97bf9ffe4-kube-api-access-z7pb2\") pod \"dnsmasq-dns-89c5cd4d5-ln75g\" (UID: \"3eed9143-556a-4df3-bef4-1eb97bf9ffe4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.640655 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.823760 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"556eac71-c652-4652-8b3e-2b208ca34704","Type":"ContainerStarted","Data":"2520cdb9ff49ebb684fd92b957bf2af995e69229d8a8ace95f101481a9e81c76"} Mar 20 16:22:07 crc kubenswrapper[4936]: I0320 16:22:07.852737 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.852697381 podStartE2EDuration="2.852697381s" podCreationTimestamp="2026-03-20 16:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:22:07.841810592 +0000 UTC m=+1278.788178407" watchObservedRunningTime="2026-03-20 16:22:07.852697381 +0000 UTC m=+1278.799065196" Mar 20 16:22:08 crc kubenswrapper[4936]: I0320 16:22:08.187976 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-ln75g"] Mar 20 16:22:08 crc kubenswrapper[4936]: I0320 16:22:08.832604 4936 generic.go:334] "Generic (PLEG): container finished" podID="3eed9143-556a-4df3-bef4-1eb97bf9ffe4" containerID="29d938d1b5e8888a85bd656252a0d5978570e2aa3ff21cfd0d0a45d0ef8cbfda" exitCode=0 Mar 20 16:22:08 crc kubenswrapper[4936]: I0320 16:22:08.832683 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" event={"ID":"3eed9143-556a-4df3-bef4-1eb97bf9ffe4","Type":"ContainerDied","Data":"29d938d1b5e8888a85bd656252a0d5978570e2aa3ff21cfd0d0a45d0ef8cbfda"} Mar 20 16:22:08 crc kubenswrapper[4936]: I0320 16:22:08.833915 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" event={"ID":"3eed9143-556a-4df3-bef4-1eb97bf9ffe4","Type":"ContainerStarted","Data":"6996c30431dd256f66b04efe6f1a5eeaf2850637d2fd2c47f2e9c2760ffd780a"} Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.421194 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.421752 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="ceilometer-central-agent" containerID="cri-o://93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37" gracePeriod=30 Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.421812 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="sg-core" containerID="cri-o://58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347" gracePeriod=30 Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.421869 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="ceilometer-notification-agent" containerID="cri-o://1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e" gracePeriod=30 Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.421877 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="proxy-httpd" containerID="cri-o://170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50" gracePeriod=30 Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.430972 4936 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.202:3000/\": read tcp 10.217.0.2:52818->10.217.0.202:3000: read: connection reset by peer" Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.848892 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" event={"ID":"3eed9143-556a-4df3-bef4-1eb97bf9ffe4","Type":"ContainerStarted","Data":"43f983d8b164558d618e46676e8c8dd689f20ce0ea179c51b5e9c9d6e37683eb"} Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.849414 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.853244 4936 generic.go:334] "Generic (PLEG): container finished" podID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerID="170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50" exitCode=0 Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.853281 4936 generic.go:334] "Generic (PLEG): container finished" podID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerID="58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347" exitCode=2 Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.886810 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" podStartSLOduration=2.886785539 podStartE2EDuration="2.886785539s" podCreationTimestamp="2026-03-20 16:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:22:09.866618226 +0000 UTC m=+1280.812986051" watchObservedRunningTime="2026-03-20 16:22:09.886785539 +0000 UTC m=+1280.833153424" Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.901989 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerDied","Data":"170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50"} Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.902055 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerDied","Data":"58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347"} Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.937001 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.937288 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-log" containerID="cri-o://6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839" gracePeriod=30 Mar 20 16:22:09 crc kubenswrapper[4936]: I0320 16:22:09.937385 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-api" containerID="cri-o://04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875" gracePeriod=30 Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.625303 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.753793 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-run-httpd\") pod \"f950892c-8013-41c7-97f2-ca160b1c00dd\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.753841 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-combined-ca-bundle\") pod \"f950892c-8013-41c7-97f2-ca160b1c00dd\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.753871 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-log-httpd\") pod \"f950892c-8013-41c7-97f2-ca160b1c00dd\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.754135 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdfrn\" (UniqueName: \"kubernetes.io/projected/f950892c-8013-41c7-97f2-ca160b1c00dd-kube-api-access-fdfrn\") pod \"f950892c-8013-41c7-97f2-ca160b1c00dd\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.754211 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-sg-core-conf-yaml\") pod \"f950892c-8013-41c7-97f2-ca160b1c00dd\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.754237 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f950892c-8013-41c7-97f2-ca160b1c00dd" (UID: "f950892c-8013-41c7-97f2-ca160b1c00dd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.754259 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-scripts\") pod \"f950892c-8013-41c7-97f2-ca160b1c00dd\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.754367 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-config-data\") pod \"f950892c-8013-41c7-97f2-ca160b1c00dd\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.754368 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f950892c-8013-41c7-97f2-ca160b1c00dd" (UID: "f950892c-8013-41c7-97f2-ca160b1c00dd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.754501 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-ceilometer-tls-certs\") pod \"f950892c-8013-41c7-97f2-ca160b1c00dd\" (UID: \"f950892c-8013-41c7-97f2-ca160b1c00dd\") " Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.755393 4936 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.755417 4936 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f950892c-8013-41c7-97f2-ca160b1c00dd-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.762862 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-scripts" (OuterVolumeSpecName: "scripts") pod "f950892c-8013-41c7-97f2-ca160b1c00dd" (UID: "f950892c-8013-41c7-97f2-ca160b1c00dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.773531 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f950892c-8013-41c7-97f2-ca160b1c00dd-kube-api-access-fdfrn" (OuterVolumeSpecName: "kube-api-access-fdfrn") pod "f950892c-8013-41c7-97f2-ca160b1c00dd" (UID: "f950892c-8013-41c7-97f2-ca160b1c00dd"). InnerVolumeSpecName "kube-api-access-fdfrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.800811 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f950892c-8013-41c7-97f2-ca160b1c00dd" (UID: "f950892c-8013-41c7-97f2-ca160b1c00dd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.817737 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f950892c-8013-41c7-97f2-ca160b1c00dd" (UID: "f950892c-8013-41c7-97f2-ca160b1c00dd"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.856727 4936 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.856760 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.856772 4936 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.856783 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdfrn\" (UniqueName: \"kubernetes.io/projected/f950892c-8013-41c7-97f2-ca160b1c00dd-kube-api-access-fdfrn\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.857314 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f950892c-8013-41c7-97f2-ca160b1c00dd" (UID: "f950892c-8013-41c7-97f2-ca160b1c00dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.868352 4936 generic.go:334] "Generic (PLEG): container finished" podID="370791fb-0e04-4268-8377-533a0544019e" containerID="6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839" exitCode=143 Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.868417 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"370791fb-0e04-4268-8377-533a0544019e","Type":"ContainerDied","Data":"6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839"} Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.875226 4936 generic.go:334] "Generic (PLEG): container finished" podID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerID="1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e" exitCode=0 Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.875262 4936 generic.go:334] "Generic (PLEG): container finished" podID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerID="93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37" exitCode=0 Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.876444 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.876560 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerDied","Data":"1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e"} Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.876594 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerDied","Data":"93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37"} Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.876609 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f950892c-8013-41c7-97f2-ca160b1c00dd","Type":"ContainerDied","Data":"8f91044be16fa217ee587605c3883ce1a268e8f09245e046e679c5953d6f0da2"} Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.876629 4936 scope.go:117] "RemoveContainer" containerID="170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.918480 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-config-data" (OuterVolumeSpecName: "config-data") pod "f950892c-8013-41c7-97f2-ca160b1c00dd" (UID: "f950892c-8013-41c7-97f2-ca160b1c00dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.920026 4936 scope.go:117] "RemoveContainer" containerID="58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.951409 4936 scope.go:117] "RemoveContainer" containerID="1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.958661 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.958690 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950892c-8013-41c7-97f2-ca160b1c00dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.973206 4936 scope.go:117] "RemoveContainer" containerID="93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.994719 4936 scope.go:117] "RemoveContainer" containerID="170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50" Mar 20 16:22:10 crc kubenswrapper[4936]: E0320 16:22:10.995186 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50\": container with ID starting with 170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50 not found: ID does not exist" containerID="170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.995219 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50"} err="failed to get container status \"170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50\": rpc error: code = NotFound desc = could not find container \"170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50\": container with ID starting with 170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50 not found: ID does not exist" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.995243 4936 scope.go:117] "RemoveContainer" containerID="58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347" Mar 20 16:22:10 crc kubenswrapper[4936]: E0320 16:22:10.995584 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347\": container with ID starting with 58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347 not found: ID does not exist" containerID="58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.995608 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347"} err="failed to get container status \"58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347\": rpc error: code = NotFound desc = could not find container \"58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347\": container with ID starting with 58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347 not found: ID does not exist" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.995622 4936 scope.go:117] "RemoveContainer" containerID="1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e" Mar 20 16:22:10 crc kubenswrapper[4936]: E0320 16:22:10.995821 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e\": container with ID starting with 1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e not found: ID does not exist" containerID="1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.995842 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e"} err="failed to get container status \"1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e\": rpc error: code = NotFound desc = could not find container \"1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e\": container with ID starting with 1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e not found: ID does not exist" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.995854 4936 scope.go:117] "RemoveContainer" containerID="93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37" Mar 20 16:22:10 crc kubenswrapper[4936]: E0320 16:22:10.996117 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37\": container with ID starting with 93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37 not found: ID does not exist" containerID="93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.996139 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37"} err="failed to get container status \"93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37\": rpc error: code = NotFound desc = could not find container \"93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37\": container with ID starting with 93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37 not found: ID does not exist" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.996151 4936 scope.go:117] "RemoveContainer" containerID="170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.996411 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50"} err="failed to get container status \"170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50\": rpc error: code = NotFound desc = could not find container \"170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50\": container with ID starting with 170ead53b24f23ca3f8b109c02b7a4d704f9c2778f32e8fb2270752e03ba2c50 not found: ID does not exist" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.996434 4936 scope.go:117] "RemoveContainer" containerID="58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.996643 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347"} err="failed to get container status \"58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347\": rpc error: code = NotFound desc = could not find container \"58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347\": container with ID starting with 58c0310835c0b167d874ef8412e5910ca8ef800af5ffce1e563f3458d0453347 not found: ID does not exist" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.996659 4936 scope.go:117] "RemoveContainer" containerID="1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.996813 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e"} err="failed to get container status \"1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e\": rpc error: code = NotFound desc = could not find container \"1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e\": container with ID starting with 1d4fe9e9d692aef1b0aceb7e4885f4b011447aa4c1b028525a2d93b57da7eb3e not found: ID does not exist" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.996830 4936 scope.go:117] "RemoveContainer" containerID="93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37" Mar 20 16:22:10 crc kubenswrapper[4936]: I0320 16:22:10.997604 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37"} err="failed to get container status \"93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37\": rpc error: code = NotFound desc = could not find container \"93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37\": container with ID starting with 93f6fb678198fe552ec81e04a025a8917eeb9854ca0d83aabaf37e7b480a5a37 not found: ID does not exist" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.164405 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.209572 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.218919 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.235588 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:11 crc kubenswrapper[4936]: E0320 16:22:11.235950 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="ceilometer-central-agent" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.235968 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="ceilometer-central-agent" Mar 20 16:22:11 crc kubenswrapper[4936]: E0320 16:22:11.235981 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="sg-core" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.235987 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="sg-core" Mar 20 16:22:11 crc kubenswrapper[4936]: E0320 16:22:11.236016 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="ceilometer-notification-agent" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.236026 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="ceilometer-notification-agent" Mar 20 16:22:11 crc kubenswrapper[4936]: E0320 16:22:11.236039 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="proxy-httpd" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.236045 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="proxy-httpd" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.236205 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="ceilometer-central-agent" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.236224 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="ceilometer-notification-agent" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.236240 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="proxy-httpd" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.236250 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" containerName="sg-core" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.237839 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.241998 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.242197 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.242601 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.260002 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.334901 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:11 crc kubenswrapper[4936]: E0320 16:22:11.335319 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-lzkcf log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.364380 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-scripts\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.364470 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.364631 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzkcf\" (UniqueName: \"kubernetes.io/projected/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-kube-api-access-lzkcf\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.364682 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.364708 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-run-httpd\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.364732 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.364782 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-log-httpd\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.364840 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-config-data\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.466968 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzkcf\" (UniqueName: \"kubernetes.io/projected/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-kube-api-access-lzkcf\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.467038 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.467069 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-run-httpd\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.467093 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.467135 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-log-httpd\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.467158 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-config-data\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.467205 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-scripts\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.467231 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.467906 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-log-httpd\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.468172 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-run-httpd\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.473040 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-scripts\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.473523 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-config-data\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.473562 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.474128 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.487455 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.492081 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzkcf\" (UniqueName: \"kubernetes.io/projected/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-kube-api-access-lzkcf\") pod \"ceilometer-0\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.864829 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f950892c-8013-41c7-97f2-ca160b1c00dd" path="/var/lib/kubelet/pods/f950892c-8013-41c7-97f2-ca160b1c00dd/volumes" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.886076 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.898497 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.977463 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-combined-ca-bundle\") pod \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.977889 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-ceilometer-tls-certs\") pod \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.977974 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-run-httpd\") pod \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.978008 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-log-httpd\") pod \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.978175 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-config-data\") pod \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.978266 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-sg-core-conf-yaml\") pod \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.978328 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-scripts\") pod \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.978405 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" (UID: "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.978414 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" (UID: "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.978425 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzkcf\" (UniqueName: \"kubernetes.io/projected/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-kube-api-access-lzkcf\") pod \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\" (UID: \"aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca\") " Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.979607 4936 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.979629 4936 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.982510 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-config-data" (OuterVolumeSpecName: "config-data") pod "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" (UID: "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.983292 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" (UID: "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.983822 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-scripts" (OuterVolumeSpecName: "scripts") pod "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" (UID: "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.983874 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-kube-api-access-lzkcf" (OuterVolumeSpecName: "kube-api-access-lzkcf") pod "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" (UID: "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca"). InnerVolumeSpecName "kube-api-access-lzkcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.983878 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" (UID: "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:11 crc kubenswrapper[4936]: I0320 16:22:11.984987 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" (UID: "aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:12 crc kubenswrapper[4936]: I0320 16:22:12.081691 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:12 crc kubenswrapper[4936]: I0320 16:22:12.081744 4936 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:12 crc kubenswrapper[4936]: I0320 16:22:12.081758 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:12 crc kubenswrapper[4936]: I0320 16:22:12.081770 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzkcf\" (UniqueName: \"kubernetes.io/projected/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-kube-api-access-lzkcf\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:12 crc kubenswrapper[4936]: I0320 16:22:12.081780 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:12 crc kubenswrapper[4936]: I0320 16:22:12.081790 4936 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:12 crc kubenswrapper[4936]: I0320 16:22:12.894063 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:22:12 crc kubenswrapper[4936]: I0320 16:22:12.988568 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:12 crc kubenswrapper[4936]: I0320 16:22:12.997128 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.005045 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.007392 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.009660 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.010055 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.010209 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.013800 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.102806 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djkkp\" (UniqueName: \"kubernetes.io/projected/e03d33f8-2aec-41af-8c22-48b41893d7b9-kube-api-access-djkkp\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.103097 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-scripts\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.103228 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e03d33f8-2aec-41af-8c22-48b41893d7b9-log-httpd\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.103319 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.103414 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.103517 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e03d33f8-2aec-41af-8c22-48b41893d7b9-run-httpd\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.103628 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-config-data\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.103792 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.205834 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.205922 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djkkp\" (UniqueName: \"kubernetes.io/projected/e03d33f8-2aec-41af-8c22-48b41893d7b9-kube-api-access-djkkp\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.205972 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-scripts\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.206025 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e03d33f8-2aec-41af-8c22-48b41893d7b9-log-httpd\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.206053 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.206108 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.206131 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e03d33f8-2aec-41af-8c22-48b41893d7b9-run-httpd\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.206152 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-config-data\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.206811 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e03d33f8-2aec-41af-8c22-48b41893d7b9-log-httpd\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.207333 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e03d33f8-2aec-41af-8c22-48b41893d7b9-run-httpd\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.210879 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.211071 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.211767 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-config-data\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.212426 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.212676 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e03d33f8-2aec-41af-8c22-48b41893d7b9-scripts\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.226408 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djkkp\" (UniqueName: \"kubernetes.io/projected/e03d33f8-2aec-41af-8c22-48b41893d7b9-kube-api-access-djkkp\") pod \"ceilometer-0\" (UID: \"e03d33f8-2aec-41af-8c22-48b41893d7b9\") " pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.330124 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.538465 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.613197 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-config-data\") pod \"370791fb-0e04-4268-8377-533a0544019e\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.613251 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/370791fb-0e04-4268-8377-533a0544019e-logs\") pod \"370791fb-0e04-4268-8377-533a0544019e\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.613460 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rvhg\" (UniqueName: \"kubernetes.io/projected/370791fb-0e04-4268-8377-533a0544019e-kube-api-access-4rvhg\") pod \"370791fb-0e04-4268-8377-533a0544019e\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.613489 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-combined-ca-bundle\") pod \"370791fb-0e04-4268-8377-533a0544019e\" (UID: \"370791fb-0e04-4268-8377-533a0544019e\") " Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.614302 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/370791fb-0e04-4268-8377-533a0544019e-logs" (OuterVolumeSpecName: "logs") pod "370791fb-0e04-4268-8377-533a0544019e" (UID: "370791fb-0e04-4268-8377-533a0544019e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.620611 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/370791fb-0e04-4268-8377-533a0544019e-kube-api-access-4rvhg" (OuterVolumeSpecName: "kube-api-access-4rvhg") pod "370791fb-0e04-4268-8377-533a0544019e" (UID: "370791fb-0e04-4268-8377-533a0544019e"). InnerVolumeSpecName "kube-api-access-4rvhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.646796 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-config-data" (OuterVolumeSpecName: "config-data") pod "370791fb-0e04-4268-8377-533a0544019e" (UID: "370791fb-0e04-4268-8377-533a0544019e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.647582 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "370791fb-0e04-4268-8377-533a0544019e" (UID: "370791fb-0e04-4268-8377-533a0544019e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.715764 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rvhg\" (UniqueName: \"kubernetes.io/projected/370791fb-0e04-4268-8377-533a0544019e-kube-api-access-4rvhg\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.715812 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.715822 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/370791fb-0e04-4268-8377-533a0544019e-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.715831 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/370791fb-0e04-4268-8377-533a0544019e-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:13 crc kubenswrapper[4936]: W0320 16:22:13.811442 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode03d33f8_2aec_41af_8c22_48b41893d7b9.slice/crio-dcc93a20abfaec633afd61fdb03a58d9616110882732431e08f0714787ec6b35 WatchSource:0}: Error finding container dcc93a20abfaec633afd61fdb03a58d9616110882732431e08f0714787ec6b35: Status 404 returned error can't find the container with id dcc93a20abfaec633afd61fdb03a58d9616110882732431e08f0714787ec6b35 Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.811578 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.870931 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca" path="/var/lib/kubelet/pods/aaaea6a3-e4b7-4e0c-9a19-9d11533b13ca/volumes" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.908639 4936 generic.go:334] "Generic (PLEG): container finished" podID="370791fb-0e04-4268-8377-533a0544019e" containerID="04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875" exitCode=0 Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.908698 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"370791fb-0e04-4268-8377-533a0544019e","Type":"ContainerDied","Data":"04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875"} Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.908725 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"370791fb-0e04-4268-8377-533a0544019e","Type":"ContainerDied","Data":"686c28a3ff11422dd0e209d5fb261d5e27cd844ce992c73f85a1dc5a1b5a3a26"} Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.908742 4936 scope.go:117] "RemoveContainer" containerID="04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.908852 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.915088 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e03d33f8-2aec-41af-8c22-48b41893d7b9","Type":"ContainerStarted","Data":"dcc93a20abfaec633afd61fdb03a58d9616110882732431e08f0714787ec6b35"} Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.940256 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.958451 4936 scope.go:117] "RemoveContainer" containerID="6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.962291 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.973444 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:13 crc kubenswrapper[4936]: E0320 16:22:13.974180 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-log" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.974200 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-log" Mar 20 16:22:13 crc kubenswrapper[4936]: E0320 16:22:13.974209 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-api" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.974215 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-api" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.974402 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-log" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.974414 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="370791fb-0e04-4268-8377-533a0544019e" containerName="nova-api-api" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.975416 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.978305 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.981668 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.981813 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.982626 4936 scope.go:117] "RemoveContainer" containerID="04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.983062 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:13 crc kubenswrapper[4936]: E0320 16:22:13.984641 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875\": container with ID starting with 04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875 not found: ID does not exist" containerID="04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.984690 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875"} err="failed to get container status \"04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875\": rpc error: code = NotFound desc = could not find container \"04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875\": container with ID starting with 04daa4ab0609929b140d30f18af83d8917eb7aee1525151bd8fc7bf9f4b37875 not found: ID does not exist" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.984719 4936 scope.go:117] "RemoveContainer" containerID="6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839" Mar 20 16:22:13 crc kubenswrapper[4936]: E0320 16:22:13.987400 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839\": container with ID starting with 6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839 not found: ID does not exist" containerID="6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839" Mar 20 16:22:13 crc kubenswrapper[4936]: I0320 16:22:13.987446 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839"} err="failed to get container status \"6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839\": rpc error: code = NotFound desc = could not find container \"6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839\": container with ID starting with 6a7b02b4bc1a96622733c90dde0d494a56739f7e5d03e095558c935ea3c47839 not found: ID does not exist" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.020206 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.020261 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34426d2a-73e7-42a7-b977-6c6ab39112d7-logs\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.020290 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-config-data\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.020386 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-public-tls-certs\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.020423 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/34426d2a-73e7-42a7-b977-6c6ab39112d7-kube-api-access-xgpl2\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.020490 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.121890 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.121983 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34426d2a-73e7-42a7-b977-6c6ab39112d7-logs\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.122012 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-config-data\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.122065 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-public-tls-certs\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.122092 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/34426d2a-73e7-42a7-b977-6c6ab39112d7-kube-api-access-xgpl2\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.122776 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.123042 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34426d2a-73e7-42a7-b977-6c6ab39112d7-logs\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.126930 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.127037 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-config-data\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.127201 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-public-tls-certs\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.128248 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.142918 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/34426d2a-73e7-42a7-b977-6c6ab39112d7-kube-api-access-xgpl2\") pod \"nova-api-0\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.293907 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.808258 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.929172 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e03d33f8-2aec-41af-8c22-48b41893d7b9","Type":"ContainerStarted","Data":"56763039d49abf21b89fce10c070b16fb3e354c14f3804e4c572aab8fb4b74bb"} Mar 20 16:22:14 crc kubenswrapper[4936]: I0320 16:22:14.931421 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34426d2a-73e7-42a7-b977-6c6ab39112d7","Type":"ContainerStarted","Data":"6d04cb3bed9cad9dc490c2b5be5d57944920cd5f53cdea3042ad810ec7ab03dc"} Mar 20 16:22:15 crc kubenswrapper[4936]: I0320 16:22:15.871387 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="370791fb-0e04-4268-8377-533a0544019e" path="/var/lib/kubelet/pods/370791fb-0e04-4268-8377-533a0544019e/volumes" Mar 20 16:22:15 crc kubenswrapper[4936]: I0320 16:22:15.941445 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e03d33f8-2aec-41af-8c22-48b41893d7b9","Type":"ContainerStarted","Data":"55b40cae7aaf22ab5ff25bff6abdc17a1bd38c924b49c994ee414c982951baec"} Mar 20 16:22:15 crc kubenswrapper[4936]: I0320 16:22:15.942894 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34426d2a-73e7-42a7-b977-6c6ab39112d7","Type":"ContainerStarted","Data":"2c852a89f15d6bae765ba3b060cc03050eb7b0739b99a9142337bfcf4d29fcf8"} Mar 20 16:22:15 crc kubenswrapper[4936]: I0320 16:22:15.942929 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34426d2a-73e7-42a7-b977-6c6ab39112d7","Type":"ContainerStarted","Data":"2dba5e8789ecf62a33b9261c2edebf8080fa8d60fdb009c204413c0c00359a3a"} Mar 20 16:22:15 crc kubenswrapper[4936]: I0320 16:22:15.963788 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.963771656 podStartE2EDuration="2.963771656s" podCreationTimestamp="2026-03-20 16:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:22:15.962290094 +0000 UTC m=+1286.908657919" watchObservedRunningTime="2026-03-20 16:22:15.963771656 +0000 UTC m=+1286.910139471" Mar 20 16:22:16 crc kubenswrapper[4936]: I0320 16:22:16.164375 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:16 crc kubenswrapper[4936]: I0320 16:22:16.183896 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:16 crc kubenswrapper[4936]: I0320 16:22:16.954705 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e03d33f8-2aec-41af-8c22-48b41893d7b9","Type":"ContainerStarted","Data":"1e6c1fae297d995f9c7c6f1472b9e5aac1f0714227d5ed939ff10ae3def6f69c"} Mar 20 16:22:16 crc kubenswrapper[4936]: I0320 16:22:16.972184 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.115148 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-snhvn"] Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.117201 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.120113 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.120369 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.124441 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-snhvn"] Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.178872 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.178930 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-config-data\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.178969 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-scripts\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.179285 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmkk2\" (UniqueName: \"kubernetes.io/projected/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-kube-api-access-bmkk2\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.281491 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.281684 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-config-data\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.281737 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-scripts\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.281839 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmkk2\" (UniqueName: \"kubernetes.io/projected/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-kube-api-access-bmkk2\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.289702 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.290031 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-config-data\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.290299 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-scripts\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.317286 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmkk2\" (UniqueName: \"kubernetes.io/projected/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-kube-api-access-bmkk2\") pod \"nova-cell1-cell-mapping-snhvn\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.443630 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.645587 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-ln75g" Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.773802 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-8b4pt"] Mar 20 16:22:17 crc kubenswrapper[4936]: I0320 16:22:17.774033 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" podUID="a481e8ea-35e6-4ad7-8fea-43226787831b" containerName="dnsmasq-dns" containerID="cri-o://098e130da0400ff88cfeb0c131f7492751e7cf87364c27359e6498c43b285422" gracePeriod=10 Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.007353 4936 generic.go:334] "Generic (PLEG): container finished" podID="a481e8ea-35e6-4ad7-8fea-43226787831b" containerID="098e130da0400ff88cfeb0c131f7492751e7cf87364c27359e6498c43b285422" exitCode=0 Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.008674 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" event={"ID":"a481e8ea-35e6-4ad7-8fea-43226787831b","Type":"ContainerDied","Data":"098e130da0400ff88cfeb0c131f7492751e7cf87364c27359e6498c43b285422"} Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.015112 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-snhvn"] Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.364820 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.430752 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-config\") pod \"a481e8ea-35e6-4ad7-8fea-43226787831b\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.430916 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-swift-storage-0\") pod \"a481e8ea-35e6-4ad7-8fea-43226787831b\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.430937 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-sb\") pod \"a481e8ea-35e6-4ad7-8fea-43226787831b\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.431004 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-svc\") pod \"a481e8ea-35e6-4ad7-8fea-43226787831b\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.431044 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-nb\") pod \"a481e8ea-35e6-4ad7-8fea-43226787831b\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.431114 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pzjs\" (UniqueName: \"kubernetes.io/projected/a481e8ea-35e6-4ad7-8fea-43226787831b-kube-api-access-2pzjs\") pod \"a481e8ea-35e6-4ad7-8fea-43226787831b\" (UID: \"a481e8ea-35e6-4ad7-8fea-43226787831b\") " Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.442826 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a481e8ea-35e6-4ad7-8fea-43226787831b-kube-api-access-2pzjs" (OuterVolumeSpecName: "kube-api-access-2pzjs") pod "a481e8ea-35e6-4ad7-8fea-43226787831b" (UID: "a481e8ea-35e6-4ad7-8fea-43226787831b"). InnerVolumeSpecName "kube-api-access-2pzjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.497328 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a481e8ea-35e6-4ad7-8fea-43226787831b" (UID: "a481e8ea-35e6-4ad7-8fea-43226787831b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.509991 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-config" (OuterVolumeSpecName: "config") pod "a481e8ea-35e6-4ad7-8fea-43226787831b" (UID: "a481e8ea-35e6-4ad7-8fea-43226787831b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.514322 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a481e8ea-35e6-4ad7-8fea-43226787831b" (UID: "a481e8ea-35e6-4ad7-8fea-43226787831b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.523979 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a481e8ea-35e6-4ad7-8fea-43226787831b" (UID: "a481e8ea-35e6-4ad7-8fea-43226787831b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.533462 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pzjs\" (UniqueName: \"kubernetes.io/projected/a481e8ea-35e6-4ad7-8fea-43226787831b-kube-api-access-2pzjs\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.533511 4936 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-config\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.533523 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.533534 4936 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.533561 4936 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.536367 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a481e8ea-35e6-4ad7-8fea-43226787831b" (UID: "a481e8ea-35e6-4ad7-8fea-43226787831b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:22:18 crc kubenswrapper[4936]: I0320 16:22:18.634992 4936 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a481e8ea-35e6-4ad7-8fea-43226787831b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.019878 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-snhvn" event={"ID":"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c","Type":"ContainerStarted","Data":"3ae29d5612c535bdbba0cdff24bd5f0a1c04d915d4d35c321b334d64d27e1ae5"} Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.019955 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-snhvn" event={"ID":"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c","Type":"ContainerStarted","Data":"68908088bfca907c860b4df295b54ea4e783d0fae64c79fd35c50e1a481e5cbc"} Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.022349 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e03d33f8-2aec-41af-8c22-48b41893d7b9","Type":"ContainerStarted","Data":"f11eb8f033c8cbdf00124d8c20fd10fc696f71868d9d0d3185eae4f6f1ea06e8"} Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.022521 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.029642 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" event={"ID":"a481e8ea-35e6-4ad7-8fea-43226787831b","Type":"ContainerDied","Data":"491708855fa1d4410b7655b256d92ae9a600e6d88173340c99c7f246387d90c4"} Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.029696 4936 scope.go:117] "RemoveContainer" containerID="098e130da0400ff88cfeb0c131f7492751e7cf87364c27359e6498c43b285422" Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.029795 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-8b4pt" Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.043843 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-snhvn" podStartSLOduration=2.043821526 podStartE2EDuration="2.043821526s" podCreationTimestamp="2026-03-20 16:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:22:19.039244515 +0000 UTC m=+1289.985612330" watchObservedRunningTime="2026-03-20 16:22:19.043821526 +0000 UTC m=+1289.990189341" Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.074388 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.11115064 podStartE2EDuration="7.074371104s" podCreationTimestamp="2026-03-20 16:22:12 +0000 UTC" firstStartedPulling="2026-03-20 16:22:13.814844273 +0000 UTC m=+1284.761212088" lastFinishedPulling="2026-03-20 16:22:17.778064737 +0000 UTC m=+1288.724432552" observedRunningTime="2026-03-20 16:22:19.072659715 +0000 UTC m=+1290.019027560" watchObservedRunningTime="2026-03-20 16:22:19.074371104 +0000 UTC m=+1290.020738919" Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.083676 4936 scope.go:117] "RemoveContainer" containerID="ad78e2b5994b670e6484af03a854dce490cf809b038545ce04d361a96b4cc354" Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.122172 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-8b4pt"] Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.127950 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-8b4pt"] Mar 20 16:22:19 crc kubenswrapper[4936]: I0320 16:22:19.872094 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a481e8ea-35e6-4ad7-8fea-43226787831b" path="/var/lib/kubelet/pods/a481e8ea-35e6-4ad7-8fea-43226787831b/volumes" Mar 20 16:22:24 crc kubenswrapper[4936]: I0320 16:22:24.078321 4936 generic.go:334] "Generic (PLEG): container finished" podID="3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c" containerID="3ae29d5612c535bdbba0cdff24bd5f0a1c04d915d4d35c321b334d64d27e1ae5" exitCode=0 Mar 20 16:22:24 crc kubenswrapper[4936]: I0320 16:22:24.078381 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-snhvn" event={"ID":"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c","Type":"ContainerDied","Data":"3ae29d5612c535bdbba0cdff24bd5f0a1c04d915d4d35c321b334d64d27e1ae5"} Mar 20 16:22:24 crc kubenswrapper[4936]: I0320 16:22:24.294781 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 20 16:22:24 crc kubenswrapper[4936]: I0320 16:22:24.295054 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.312729 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.312756 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.493177 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.588313 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-scripts\") pod \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.588655 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-combined-ca-bundle\") pod \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.588699 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-config-data\") pod \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.588764 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmkk2\" (UniqueName: \"kubernetes.io/projected/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-kube-api-access-bmkk2\") pod \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\" (UID: \"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c\") " Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.605825 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-scripts" (OuterVolumeSpecName: "scripts") pod "3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c" (UID: "3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.606238 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-kube-api-access-bmkk2" (OuterVolumeSpecName: "kube-api-access-bmkk2") pod "3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c" (UID: "3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c"). InnerVolumeSpecName "kube-api-access-bmkk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.622621 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c" (UID: "3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.624042 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-config-data" (OuterVolumeSpecName: "config-data") pod "3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c" (UID: "3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.691125 4936 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-scripts\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.691176 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.691186 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:25 crc kubenswrapper[4936]: I0320 16:22:25.691196 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmkk2\" (UniqueName: \"kubernetes.io/projected/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c-kube-api-access-bmkk2\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.098361 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-snhvn" event={"ID":"3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c","Type":"ContainerDied","Data":"68908088bfca907c860b4df295b54ea4e783d0fae64c79fd35c50e1a481e5cbc"} Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.098410 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68908088bfca907c860b4df295b54ea4e783d0fae64c79fd35c50e1a481e5cbc" Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.098434 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-snhvn" Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.289881 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.290171 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="265ab68a-88f2-47c2-88d6-9afe238f1733" containerName="nova-scheduler-scheduler" containerID="cri-o://67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6" gracePeriod=30 Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.301323 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.301630 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-log" containerID="cri-o://2dba5e8789ecf62a33b9261c2edebf8080fa8d60fdb009c204413c0c00359a3a" gracePeriod=30 Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.301699 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-api" containerID="cri-o://2c852a89f15d6bae765ba3b060cc03050eb7b0739b99a9142337bfcf4d29fcf8" gracePeriod=30 Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.319152 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.319403 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-log" containerID="cri-o://84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de" gracePeriod=30 Mar 20 16:22:26 crc kubenswrapper[4936]: I0320 16:22:26.319519 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-metadata" containerID="cri-o://8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708" gracePeriod=30 Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.109755 4936 generic.go:334] "Generic (PLEG): container finished" podID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerID="84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de" exitCode=143 Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.109829 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4180e1-8bc4-4b72-9d32-323b920e28e3","Type":"ContainerDied","Data":"84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de"} Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.112209 4936 generic.go:334] "Generic (PLEG): container finished" podID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerID="2dba5e8789ecf62a33b9261c2edebf8080fa8d60fdb009c204413c0c00359a3a" exitCode=143 Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.112232 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34426d2a-73e7-42a7-b977-6c6ab39112d7","Type":"ContainerDied","Data":"2dba5e8789ecf62a33b9261c2edebf8080fa8d60fdb009c204413c0c00359a3a"} Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.793466 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.836062 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sq75\" (UniqueName: \"kubernetes.io/projected/265ab68a-88f2-47c2-88d6-9afe238f1733-kube-api-access-6sq75\") pod \"265ab68a-88f2-47c2-88d6-9afe238f1733\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.836239 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-combined-ca-bundle\") pod \"265ab68a-88f2-47c2-88d6-9afe238f1733\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.836259 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-config-data\") pod \"265ab68a-88f2-47c2-88d6-9afe238f1733\" (UID: \"265ab68a-88f2-47c2-88d6-9afe238f1733\") " Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.857885 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/265ab68a-88f2-47c2-88d6-9afe238f1733-kube-api-access-6sq75" (OuterVolumeSpecName: "kube-api-access-6sq75") pod "265ab68a-88f2-47c2-88d6-9afe238f1733" (UID: "265ab68a-88f2-47c2-88d6-9afe238f1733"). InnerVolumeSpecName "kube-api-access-6sq75". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.869802 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-config-data" (OuterVolumeSpecName: "config-data") pod "265ab68a-88f2-47c2-88d6-9afe238f1733" (UID: "265ab68a-88f2-47c2-88d6-9afe238f1733"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.880718 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "265ab68a-88f2-47c2-88d6-9afe238f1733" (UID: "265ab68a-88f2-47c2-88d6-9afe238f1733"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.938127 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.938166 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/265ab68a-88f2-47c2-88d6-9afe238f1733-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:27 crc kubenswrapper[4936]: I0320 16:22:27.938178 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sq75\" (UniqueName: \"kubernetes.io/projected/265ab68a-88f2-47c2-88d6-9afe238f1733-kube-api-access-6sq75\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.126386 4936 generic.go:334] "Generic (PLEG): container finished" podID="265ab68a-88f2-47c2-88d6-9afe238f1733" containerID="67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6" exitCode=0 Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.126441 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"265ab68a-88f2-47c2-88d6-9afe238f1733","Type":"ContainerDied","Data":"67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6"} Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.126475 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"265ab68a-88f2-47c2-88d6-9afe238f1733","Type":"ContainerDied","Data":"a589bd3309476876c16fb47df8674cddcc00542217e31c3baa0722d9512fb5ba"} Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.126492 4936 scope.go:117] "RemoveContainer" containerID="67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.126692 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.178281 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.183659 4936 scope.go:117] "RemoveContainer" containerID="67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6" Mar 20 16:22:28 crc kubenswrapper[4936]: E0320 16:22:28.192288 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6\": container with ID starting with 67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6 not found: ID does not exist" containerID="67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.192652 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6"} err="failed to get container status \"67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6\": rpc error: code = NotFound desc = could not find container \"67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6\": container with ID starting with 67592e3fcad1223f81bc01d815b1fe73cc23556807b75c6dd3a45eb1ae92a8f6 not found: ID does not exist" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.196386 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.205510 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:22:28 crc kubenswrapper[4936]: E0320 16:22:28.205985 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a481e8ea-35e6-4ad7-8fea-43226787831b" containerName="dnsmasq-dns" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.206008 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a481e8ea-35e6-4ad7-8fea-43226787831b" containerName="dnsmasq-dns" Mar 20 16:22:28 crc kubenswrapper[4936]: E0320 16:22:28.206043 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c" containerName="nova-manage" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.206072 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c" containerName="nova-manage" Mar 20 16:22:28 crc kubenswrapper[4936]: E0320 16:22:28.206085 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="265ab68a-88f2-47c2-88d6-9afe238f1733" containerName="nova-scheduler-scheduler" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.206093 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="265ab68a-88f2-47c2-88d6-9afe238f1733" containerName="nova-scheduler-scheduler" Mar 20 16:22:28 crc kubenswrapper[4936]: E0320 16:22:28.206109 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a481e8ea-35e6-4ad7-8fea-43226787831b" containerName="init" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.206116 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="a481e8ea-35e6-4ad7-8fea-43226787831b" containerName="init" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.206359 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c" containerName="nova-manage" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.206408 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="265ab68a-88f2-47c2-88d6-9afe238f1733" containerName="nova-scheduler-scheduler" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.206427 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="a481e8ea-35e6-4ad7-8fea-43226787831b" containerName="dnsmasq-dns" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.207231 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.209213 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.216672 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.344278 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz4wh\" (UniqueName: \"kubernetes.io/projected/e3bf3212-25ad-4ec7-bb5d-c162c5428220-kube-api-access-xz4wh\") pod \"nova-scheduler-0\" (UID: \"e3bf3212-25ad-4ec7-bb5d-c162c5428220\") " pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.344471 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3bf3212-25ad-4ec7-bb5d-c162c5428220-config-data\") pod \"nova-scheduler-0\" (UID: \"e3bf3212-25ad-4ec7-bb5d-c162c5428220\") " pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.344633 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3bf3212-25ad-4ec7-bb5d-c162c5428220-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e3bf3212-25ad-4ec7-bb5d-c162c5428220\") " pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.446177 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3bf3212-25ad-4ec7-bb5d-c162c5428220-config-data\") pod \"nova-scheduler-0\" (UID: \"e3bf3212-25ad-4ec7-bb5d-c162c5428220\") " pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.446256 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3bf3212-25ad-4ec7-bb5d-c162c5428220-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e3bf3212-25ad-4ec7-bb5d-c162c5428220\") " pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.446339 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz4wh\" (UniqueName: \"kubernetes.io/projected/e3bf3212-25ad-4ec7-bb5d-c162c5428220-kube-api-access-xz4wh\") pod \"nova-scheduler-0\" (UID: \"e3bf3212-25ad-4ec7-bb5d-c162c5428220\") " pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.451053 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3bf3212-25ad-4ec7-bb5d-c162c5428220-config-data\") pod \"nova-scheduler-0\" (UID: \"e3bf3212-25ad-4ec7-bb5d-c162c5428220\") " pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.452657 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3bf3212-25ad-4ec7-bb5d-c162c5428220-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e3bf3212-25ad-4ec7-bb5d-c162c5428220\") " pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.468078 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz4wh\" (UniqueName: \"kubernetes.io/projected/e3bf3212-25ad-4ec7-bb5d-c162c5428220-kube-api-access-xz4wh\") pod \"nova-scheduler-0\" (UID: \"e3bf3212-25ad-4ec7-bb5d-c162c5428220\") " pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.537495 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.854620 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.854937 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:22:28 crc kubenswrapper[4936]: I0320 16:22:28.981764 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 20 16:22:29 crc kubenswrapper[4936]: I0320 16:22:29.136813 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e3bf3212-25ad-4ec7-bb5d-c162c5428220","Type":"ContainerStarted","Data":"91aea791ed054746086a6067ac22678276c473b62cbd6be9835044421de0fe3f"} Mar 20 16:22:29 crc kubenswrapper[4936]: I0320 16:22:29.868930 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="265ab68a-88f2-47c2-88d6-9afe238f1733" path="/var/lib/kubelet/pods/265ab68a-88f2-47c2-88d6-9afe238f1733/volumes" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.019605 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.076680 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-nova-metadata-tls-certs\") pod \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.076821 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-combined-ca-bundle\") pod \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.077469 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngjf5\" (UniqueName: \"kubernetes.io/projected/1a4180e1-8bc4-4b72-9d32-323b920e28e3-kube-api-access-ngjf5\") pod \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.077520 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-config-data\") pod \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.077665 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4180e1-8bc4-4b72-9d32-323b920e28e3-logs\") pod \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\" (UID: \"1a4180e1-8bc4-4b72-9d32-323b920e28e3\") " Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.078450 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a4180e1-8bc4-4b72-9d32-323b920e28e3-logs" (OuterVolumeSpecName: "logs") pod "1a4180e1-8bc4-4b72-9d32-323b920e28e3" (UID: "1a4180e1-8bc4-4b72-9d32-323b920e28e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.081723 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a4180e1-8bc4-4b72-9d32-323b920e28e3-kube-api-access-ngjf5" (OuterVolumeSpecName: "kube-api-access-ngjf5") pod "1a4180e1-8bc4-4b72-9d32-323b920e28e3" (UID: "1a4180e1-8bc4-4b72-9d32-323b920e28e3"). InnerVolumeSpecName "kube-api-access-ngjf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.105170 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a4180e1-8bc4-4b72-9d32-323b920e28e3" (UID: "1a4180e1-8bc4-4b72-9d32-323b920e28e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.127775 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-config-data" (OuterVolumeSpecName: "config-data") pod "1a4180e1-8bc4-4b72-9d32-323b920e28e3" (UID: "1a4180e1-8bc4-4b72-9d32-323b920e28e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.148015 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e3bf3212-25ad-4ec7-bb5d-c162c5428220","Type":"ContainerStarted","Data":"39d5d89d5c929a05db672b9951cf8c7cbab5cdba522231c0b293aec415a6dc1c"} Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.150363 4936 generic.go:334] "Generic (PLEG): container finished" podID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerID="8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708" exitCode=0 Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.150406 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4180e1-8bc4-4b72-9d32-323b920e28e3","Type":"ContainerDied","Data":"8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708"} Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.150432 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4180e1-8bc4-4b72-9d32-323b920e28e3","Type":"ContainerDied","Data":"f6ec3f795c9d42c4988347fce54278a7badfd54a4c45e9463d81fdd697b68761"} Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.150452 4936 scope.go:117] "RemoveContainer" containerID="8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.150555 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.157761 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1a4180e1-8bc4-4b72-9d32-323b920e28e3" (UID: "1a4180e1-8bc4-4b72-9d32-323b920e28e3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.179261 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.179239958 podStartE2EDuration="2.179239958s" podCreationTimestamp="2026-03-20 16:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:22:30.170197601 +0000 UTC m=+1301.116565416" watchObservedRunningTime="2026-03-20 16:22:30.179239958 +0000 UTC m=+1301.125607773" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.180795 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4180e1-8bc4-4b72-9d32-323b920e28e3-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.181440 4936 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.181492 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.181509 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngjf5\" (UniqueName: \"kubernetes.io/projected/1a4180e1-8bc4-4b72-9d32-323b920e28e3-kube-api-access-ngjf5\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.181595 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4180e1-8bc4-4b72-9d32-323b920e28e3-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.281825 4936 scope.go:117] "RemoveContainer" containerID="84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.304008 4936 scope.go:117] "RemoveContainer" containerID="8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708" Mar 20 16:22:30 crc kubenswrapper[4936]: E0320 16:22:30.304462 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708\": container with ID starting with 8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708 not found: ID does not exist" containerID="8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.304490 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708"} err="failed to get container status \"8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708\": rpc error: code = NotFound desc = could not find container \"8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708\": container with ID starting with 8810c4faa8b46c814c36869c15c306acc2a27a8eba9ffe441923ef20b7159708 not found: ID does not exist" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.304512 4936 scope.go:117] "RemoveContainer" containerID="84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de" Mar 20 16:22:30 crc kubenswrapper[4936]: E0320 16:22:30.304757 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de\": container with ID starting with 84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de not found: ID does not exist" containerID="84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.304783 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de"} err="failed to get container status \"84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de\": rpc error: code = NotFound desc = could not find container \"84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de\": container with ID starting with 84bc08c5228cfc0e44bb023645b1ab84c26f450a79effbd8d97db7bc128641de not found: ID does not exist" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.506960 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.531819 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.542593 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:22:30 crc kubenswrapper[4936]: E0320 16:22:30.543103 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-metadata" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.543128 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-metadata" Mar 20 16:22:30 crc kubenswrapper[4936]: E0320 16:22:30.543162 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-log" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.543171 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-log" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.543431 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-log" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.543467 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" containerName="nova-metadata-metadata" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.544684 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.551513 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.553161 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.556971 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.590042 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e89b3de-e3f1-4258-9f7d-937b8159f213-logs\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.590500 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e89b3de-e3f1-4258-9f7d-937b8159f213-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.590636 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e89b3de-e3f1-4258-9f7d-937b8159f213-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.590786 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbqjr\" (UniqueName: \"kubernetes.io/projected/7e89b3de-e3f1-4258-9f7d-937b8159f213-kube-api-access-kbqjr\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.590952 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e89b3de-e3f1-4258-9f7d-937b8159f213-config-data\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.692768 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e89b3de-e3f1-4258-9f7d-937b8159f213-logs\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.693161 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e89b3de-e3f1-4258-9f7d-937b8159f213-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.693277 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e89b3de-e3f1-4258-9f7d-937b8159f213-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.693427 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbqjr\" (UniqueName: \"kubernetes.io/projected/7e89b3de-e3f1-4258-9f7d-937b8159f213-kube-api-access-kbqjr\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.693465 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e89b3de-e3f1-4258-9f7d-937b8159f213-logs\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.693763 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e89b3de-e3f1-4258-9f7d-937b8159f213-config-data\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.696353 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e89b3de-e3f1-4258-9f7d-937b8159f213-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.697243 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e89b3de-e3f1-4258-9f7d-937b8159f213-config-data\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.698389 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e89b3de-e3f1-4258-9f7d-937b8159f213-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.711336 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbqjr\" (UniqueName: \"kubernetes.io/projected/7e89b3de-e3f1-4258-9f7d-937b8159f213-kube-api-access-kbqjr\") pod \"nova-metadata-0\" (UID: \"7e89b3de-e3f1-4258-9f7d-937b8159f213\") " pod="openstack/nova-metadata-0" Mar 20 16:22:30 crc kubenswrapper[4936]: I0320 16:22:30.878410 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.160084 4936 generic.go:334] "Generic (PLEG): container finished" podID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerID="2c852a89f15d6bae765ba3b060cc03050eb7b0739b99a9142337bfcf4d29fcf8" exitCode=0 Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.161386 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34426d2a-73e7-42a7-b977-6c6ab39112d7","Type":"ContainerDied","Data":"2c852a89f15d6bae765ba3b060cc03050eb7b0739b99a9142337bfcf4d29fcf8"} Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.310095 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.390165 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 20 16:22:31 crc kubenswrapper[4936]: W0320 16:22:31.405163 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e89b3de_e3f1_4258_9f7d_937b8159f213.slice/crio-cc8ca8cc49f455b85108dee33629c294e3e7053078b9d29cd8da789e9b066c78 WatchSource:0}: Error finding container cc8ca8cc49f455b85108dee33629c294e3e7053078b9d29cd8da789e9b066c78: Status 404 returned error can't find the container with id cc8ca8cc49f455b85108dee33629c294e3e7053078b9d29cd8da789e9b066c78 Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.411891 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-public-tls-certs\") pod \"34426d2a-73e7-42a7-b977-6c6ab39112d7\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.412047 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/34426d2a-73e7-42a7-b977-6c6ab39112d7-kube-api-access-xgpl2\") pod \"34426d2a-73e7-42a7-b977-6c6ab39112d7\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.412096 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-combined-ca-bundle\") pod \"34426d2a-73e7-42a7-b977-6c6ab39112d7\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.412137 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34426d2a-73e7-42a7-b977-6c6ab39112d7-logs\") pod \"34426d2a-73e7-42a7-b977-6c6ab39112d7\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.412174 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-config-data\") pod \"34426d2a-73e7-42a7-b977-6c6ab39112d7\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.412263 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-internal-tls-certs\") pod \"34426d2a-73e7-42a7-b977-6c6ab39112d7\" (UID: \"34426d2a-73e7-42a7-b977-6c6ab39112d7\") " Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.414156 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34426d2a-73e7-42a7-b977-6c6ab39112d7-logs" (OuterVolumeSpecName: "logs") pod "34426d2a-73e7-42a7-b977-6c6ab39112d7" (UID: "34426d2a-73e7-42a7-b977-6c6ab39112d7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.419808 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34426d2a-73e7-42a7-b977-6c6ab39112d7-kube-api-access-xgpl2" (OuterVolumeSpecName: "kube-api-access-xgpl2") pod "34426d2a-73e7-42a7-b977-6c6ab39112d7" (UID: "34426d2a-73e7-42a7-b977-6c6ab39112d7"). InnerVolumeSpecName "kube-api-access-xgpl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.449217 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34426d2a-73e7-42a7-b977-6c6ab39112d7" (UID: "34426d2a-73e7-42a7-b977-6c6ab39112d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.475724 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-config-data" (OuterVolumeSpecName: "config-data") pod "34426d2a-73e7-42a7-b977-6c6ab39112d7" (UID: "34426d2a-73e7-42a7-b977-6c6ab39112d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.480272 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "34426d2a-73e7-42a7-b977-6c6ab39112d7" (UID: "34426d2a-73e7-42a7-b977-6c6ab39112d7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.486192 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "34426d2a-73e7-42a7-b977-6c6ab39112d7" (UID: "34426d2a-73e7-42a7-b977-6c6ab39112d7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.514858 4936 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.514892 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/34426d2a-73e7-42a7-b977-6c6ab39112d7-kube-api-access-xgpl2\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.514902 4936 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.514913 4936 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34426d2a-73e7-42a7-b977-6c6ab39112d7-logs\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.514921 4936 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-config-data\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.514929 4936 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34426d2a-73e7-42a7-b977-6c6ab39112d7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 20 16:22:31 crc kubenswrapper[4936]: I0320 16:22:31.865834 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a4180e1-8bc4-4b72-9d32-323b920e28e3" path="/var/lib/kubelet/pods/1a4180e1-8bc4-4b72-9d32-323b920e28e3/volumes" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.170809 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34426d2a-73e7-42a7-b977-6c6ab39112d7","Type":"ContainerDied","Data":"6d04cb3bed9cad9dc490c2b5be5d57944920cd5f53cdea3042ad810ec7ab03dc"} Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.171127 4936 scope.go:117] "RemoveContainer" containerID="2c852a89f15d6bae765ba3b060cc03050eb7b0739b99a9142337bfcf4d29fcf8" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.170899 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.174647 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e89b3de-e3f1-4258-9f7d-937b8159f213","Type":"ContainerStarted","Data":"8f275895ba505f71b0d26af2aeb04a5bc716fd52a656679c96e8e8afb607a230"} Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.174684 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e89b3de-e3f1-4258-9f7d-937b8159f213","Type":"ContainerStarted","Data":"121476e379588a048004fbc1253c8c5586533836ff33178fd58a159ee5081627"} Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.174693 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e89b3de-e3f1-4258-9f7d-937b8159f213","Type":"ContainerStarted","Data":"cc8ca8cc49f455b85108dee33629c294e3e7053078b9d29cd8da789e9b066c78"} Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.200664 4936 scope.go:117] "RemoveContainer" containerID="2dba5e8789ecf62a33b9261c2edebf8080fa8d60fdb009c204413c0c00359a3a" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.203297 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.2032763109999998 podStartE2EDuration="2.203276311s" podCreationTimestamp="2026-03-20 16:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:22:32.1944489 +0000 UTC m=+1303.140816725" watchObservedRunningTime="2026-03-20 16:22:32.203276311 +0000 UTC m=+1303.149644136" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.221023 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.237586 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.251406 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:32 crc kubenswrapper[4936]: E0320 16:22:32.251847 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-log" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.251867 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-log" Mar 20 16:22:32 crc kubenswrapper[4936]: E0320 16:22:32.251893 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-api" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.251900 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-api" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.252067 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-log" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.252085 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" containerName="nova-api-api" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.253011 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.261197 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.280138 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.280384 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.280645 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.337144 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-config-data\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.337245 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzs7g\" (UniqueName: \"kubernetes.io/projected/02a89fea-4ef7-41e3-a914-5cec1d6c7439-kube-api-access-gzs7g\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.337316 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-public-tls-certs\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.337350 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.337391 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02a89fea-4ef7-41e3-a914-5cec1d6c7439-logs\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.337411 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-internal-tls-certs\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.439152 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.439221 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02a89fea-4ef7-41e3-a914-5cec1d6c7439-logs\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.439256 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-internal-tls-certs\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.439377 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-config-data\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.439431 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzs7g\" (UniqueName: \"kubernetes.io/projected/02a89fea-4ef7-41e3-a914-5cec1d6c7439-kube-api-access-gzs7g\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.439504 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-public-tls-certs\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.439794 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02a89fea-4ef7-41e3-a914-5cec1d6c7439-logs\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.444321 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.444691 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-config-data\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.450854 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-public-tls-certs\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.455917 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02a89fea-4ef7-41e3-a914-5cec1d6c7439-internal-tls-certs\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.457217 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzs7g\" (UniqueName: \"kubernetes.io/projected/02a89fea-4ef7-41e3-a914-5cec1d6c7439-kube-api-access-gzs7g\") pod \"nova-api-0\" (UID: \"02a89fea-4ef7-41e3-a914-5cec1d6c7439\") " pod="openstack/nova-api-0" Mar 20 16:22:32 crc kubenswrapper[4936]: I0320 16:22:32.598826 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 20 16:22:33 crc kubenswrapper[4936]: I0320 16:22:33.057826 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 20 16:22:33 crc kubenswrapper[4936]: I0320 16:22:33.186058 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"02a89fea-4ef7-41e3-a914-5cec1d6c7439","Type":"ContainerStarted","Data":"a6b4642cca7881cb067a2cd34b250b97334be8fa14e2becddfe7f97f1dd34ac3"} Mar 20 16:22:33 crc kubenswrapper[4936]: I0320 16:22:33.538562 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 20 16:22:33 crc kubenswrapper[4936]: I0320 16:22:33.865806 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34426d2a-73e7-42a7-b977-6c6ab39112d7" path="/var/lib/kubelet/pods/34426d2a-73e7-42a7-b977-6c6ab39112d7/volumes" Mar 20 16:22:34 crc kubenswrapper[4936]: I0320 16:22:34.194853 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"02a89fea-4ef7-41e3-a914-5cec1d6c7439","Type":"ContainerStarted","Data":"c2bce3cc9f80dba89a949c5848246c87cbff3efb31b45e0d1c8af263c6708a8a"} Mar 20 16:22:34 crc kubenswrapper[4936]: I0320 16:22:34.194930 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"02a89fea-4ef7-41e3-a914-5cec1d6c7439","Type":"ContainerStarted","Data":"eb59fcb284cae838f8b4d115f9ffdb17315766e6da28626ce8e7753099784008"} Mar 20 16:22:34 crc kubenswrapper[4936]: I0320 16:22:34.225097 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.22507928 podStartE2EDuration="2.22507928s" podCreationTimestamp="2026-03-20 16:22:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 16:22:34.210423103 +0000 UTC m=+1305.156790948" watchObservedRunningTime="2026-03-20 16:22:34.22507928 +0000 UTC m=+1305.171447095" Mar 20 16:22:38 crc kubenswrapper[4936]: I0320 16:22:38.538685 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 20 16:22:38 crc kubenswrapper[4936]: I0320 16:22:38.573918 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 20 16:22:39 crc kubenswrapper[4936]: I0320 16:22:39.262141 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 20 16:22:40 crc kubenswrapper[4936]: I0320 16:22:40.585211 4936 scope.go:117] "RemoveContainer" containerID="f04149f5ca7c1e8786a41e139ac627021de4b5a3a4a6079a996d6a5f0bd416d8" Mar 20 16:22:40 crc kubenswrapper[4936]: I0320 16:22:40.879306 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 20 16:22:40 crc kubenswrapper[4936]: I0320 16:22:40.879703 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 20 16:22:41 crc kubenswrapper[4936]: I0320 16:22:41.897692 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7e89b3de-e3f1-4258-9f7d-937b8159f213" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 20 16:22:41 crc kubenswrapper[4936]: I0320 16:22:41.897778 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7e89b3de-e3f1-4258-9f7d-937b8159f213" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 20 16:22:42 crc kubenswrapper[4936]: I0320 16:22:42.599889 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 20 16:22:42 crc kubenswrapper[4936]: I0320 16:22:42.599950 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 20 16:22:43 crc kubenswrapper[4936]: I0320 16:22:43.353358 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 20 16:22:43 crc kubenswrapper[4936]: I0320 16:22:43.610891 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="02a89fea-4ef7-41e3-a914-5cec1d6c7439" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 20 16:22:43 crc kubenswrapper[4936]: I0320 16:22:43.610872 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="02a89fea-4ef7-41e3-a914-5cec1d6c7439" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 20 16:22:48 crc kubenswrapper[4936]: I0320 16:22:48.879307 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 20 16:22:48 crc kubenswrapper[4936]: I0320 16:22:48.879874 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 20 16:22:50 crc kubenswrapper[4936]: I0320 16:22:50.599475 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 20 16:22:50 crc kubenswrapper[4936]: I0320 16:22:50.600887 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 20 16:22:50 crc kubenswrapper[4936]: I0320 16:22:50.892802 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 20 16:22:50 crc kubenswrapper[4936]: I0320 16:22:50.893680 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 20 16:22:50 crc kubenswrapper[4936]: I0320 16:22:50.900259 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 20 16:22:50 crc kubenswrapper[4936]: I0320 16:22:50.902600 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 20 16:22:52 crc kubenswrapper[4936]: I0320 16:22:52.606643 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 20 16:22:52 crc kubenswrapper[4936]: I0320 16:22:52.606712 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 20 16:22:52 crc kubenswrapper[4936]: I0320 16:22:52.613415 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 20 16:22:52 crc kubenswrapper[4936]: I0320 16:22:52.613613 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 20 16:22:58 crc kubenswrapper[4936]: I0320 16:22:58.854670 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:22:58 crc kubenswrapper[4936]: I0320 16:22:58.855212 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:23:28 crc kubenswrapper[4936]: I0320 16:23:28.855186 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:23:28 crc kubenswrapper[4936]: I0320 16:23:28.855739 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:23:28 crc kubenswrapper[4936]: I0320 16:23:28.855775 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:23:28 crc kubenswrapper[4936]: I0320 16:23:28.856398 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"19b7ae40534ebc0df1ca16d7a38e73567bfe3ddd0d163640d422629548acfbe8"} pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 16:23:28 crc kubenswrapper[4936]: I0320 16:23:28.856498 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" containerID="cri-o://19b7ae40534ebc0df1ca16d7a38e73567bfe3ddd0d163640d422629548acfbe8" gracePeriod=600 Mar 20 16:23:29 crc kubenswrapper[4936]: I0320 16:23:29.709867 4936 generic.go:334] "Generic (PLEG): container finished" podID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerID="19b7ae40534ebc0df1ca16d7a38e73567bfe3ddd0d163640d422629548acfbe8" exitCode=0 Mar 20 16:23:29 crc kubenswrapper[4936]: I0320 16:23:29.709935 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerDied","Data":"19b7ae40534ebc0df1ca16d7a38e73567bfe3ddd0d163640d422629548acfbe8"} Mar 20 16:23:29 crc kubenswrapper[4936]: I0320 16:23:29.710483 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"7f54448fa33df6b9490f8d0906969969ecfdaf5a9c74fdcedd6ab26b4c513d8e"} Mar 20 16:23:29 crc kubenswrapper[4936]: I0320 16:23:29.710528 4936 scope.go:117] "RemoveContainer" containerID="929e303b68c86406c3ac9c3e153cfe435d16f382342ffc919886659461c4db6e" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.276505 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xpdzx/must-gather-glbch"] Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.280429 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.289439 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xpdzx"/"openshift-service-ca.crt" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.309820 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xpdzx"/"kube-root-ca.crt" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.332277 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xpdzx/must-gather-glbch"] Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.375213 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/631b69e4-abc4-431f-9edb-d076726545e7-must-gather-output\") pod \"must-gather-glbch\" (UID: \"631b69e4-abc4-431f-9edb-d076726545e7\") " pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.375285 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pzwj\" (UniqueName: \"kubernetes.io/projected/631b69e4-abc4-431f-9edb-d076726545e7-kube-api-access-7pzwj\") pod \"must-gather-glbch\" (UID: \"631b69e4-abc4-431f-9edb-d076726545e7\") " pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.485807 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/631b69e4-abc4-431f-9edb-d076726545e7-must-gather-output\") pod \"must-gather-glbch\" (UID: \"631b69e4-abc4-431f-9edb-d076726545e7\") " pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.485901 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pzwj\" (UniqueName: \"kubernetes.io/projected/631b69e4-abc4-431f-9edb-d076726545e7-kube-api-access-7pzwj\") pod \"must-gather-glbch\" (UID: \"631b69e4-abc4-431f-9edb-d076726545e7\") " pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.488664 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/631b69e4-abc4-431f-9edb-d076726545e7-must-gather-output\") pod \"must-gather-glbch\" (UID: \"631b69e4-abc4-431f-9edb-d076726545e7\") " pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.513650 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pzwj\" (UniqueName: \"kubernetes.io/projected/631b69e4-abc4-431f-9edb-d076726545e7-kube-api-access-7pzwj\") pod \"must-gather-glbch\" (UID: \"631b69e4-abc4-431f-9edb-d076726545e7\") " pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:23:40 crc kubenswrapper[4936]: I0320 16:23:40.615737 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:23:41 crc kubenswrapper[4936]: I0320 16:23:41.130851 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xpdzx/must-gather-glbch"] Mar 20 16:23:41 crc kubenswrapper[4936]: I0320 16:23:41.822928 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xpdzx/must-gather-glbch" event={"ID":"631b69e4-abc4-431f-9edb-d076726545e7","Type":"ContainerStarted","Data":"91517031d5e5f8b2894fec91df607bc02b54e3605226d2069bdf233a502deb78"} Mar 20 16:23:45 crc kubenswrapper[4936]: I0320 16:23:45.878618 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xpdzx/must-gather-glbch" event={"ID":"631b69e4-abc4-431f-9edb-d076726545e7","Type":"ContainerStarted","Data":"12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9"} Mar 20 16:23:45 crc kubenswrapper[4936]: I0320 16:23:45.879153 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xpdzx/must-gather-glbch" event={"ID":"631b69e4-abc4-431f-9edb-d076726545e7","Type":"ContainerStarted","Data":"937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e"} Mar 20 16:23:45 crc kubenswrapper[4936]: I0320 16:23:45.899801 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xpdzx/must-gather-glbch" podStartSLOduration=1.849786295 podStartE2EDuration="5.899780462s" podCreationTimestamp="2026-03-20 16:23:40 +0000 UTC" firstStartedPulling="2026-03-20 16:23:41.134592158 +0000 UTC m=+1372.080959973" lastFinishedPulling="2026-03-20 16:23:45.184586325 +0000 UTC m=+1376.130954140" observedRunningTime="2026-03-20 16:23:45.895201783 +0000 UTC m=+1376.841569598" watchObservedRunningTime="2026-03-20 16:23:45.899780462 +0000 UTC m=+1376.846148277" Mar 20 16:23:50 crc kubenswrapper[4936]: I0320 16:23:50.685309 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xpdzx/crc-debug-5l4z6"] Mar 20 16:23:50 crc kubenswrapper[4936]: I0320 16:23:50.687081 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:23:50 crc kubenswrapper[4936]: I0320 16:23:50.689265 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xpdzx"/"default-dockercfg-85kmq" Mar 20 16:23:50 crc kubenswrapper[4936]: I0320 16:23:50.817692 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-host\") pod \"crc-debug-5l4z6\" (UID: \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\") " pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:23:50 crc kubenswrapper[4936]: I0320 16:23:50.818253 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjmtj\" (UniqueName: \"kubernetes.io/projected/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-kube-api-access-qjmtj\") pod \"crc-debug-5l4z6\" (UID: \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\") " pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:23:50 crc kubenswrapper[4936]: I0320 16:23:50.921206 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-host\") pod \"crc-debug-5l4z6\" (UID: \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\") " pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:23:50 crc kubenswrapper[4936]: I0320 16:23:50.921436 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjmtj\" (UniqueName: \"kubernetes.io/projected/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-kube-api-access-qjmtj\") pod \"crc-debug-5l4z6\" (UID: \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\") " pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:23:50 crc kubenswrapper[4936]: I0320 16:23:50.924162 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-host\") pod \"crc-debug-5l4z6\" (UID: \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\") " pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:23:50 crc kubenswrapper[4936]: I0320 16:23:50.948450 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjmtj\" (UniqueName: \"kubernetes.io/projected/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-kube-api-access-qjmtj\") pod \"crc-debug-5l4z6\" (UID: \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\") " pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:23:51 crc kubenswrapper[4936]: I0320 16:23:51.005534 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:23:51 crc kubenswrapper[4936]: I0320 16:23:51.044481 4936 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 16:23:51 crc kubenswrapper[4936]: I0320 16:23:51.962817 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" event={"ID":"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127","Type":"ContainerStarted","Data":"fda4938a42b573233f34959f3d107e166a561ccef4cfe980b2e74e751b201a4d"} Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.161437 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567064-8kqzc"] Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.163225 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567064-8kqzc" Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.169925 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.170144 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.170338 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.200147 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567064-8kqzc"] Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.207602 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwghq\" (UniqueName: \"kubernetes.io/projected/83b71159-8ac2-4910-86ea-639368e9fe03-kube-api-access-qwghq\") pod \"auto-csr-approver-29567064-8kqzc\" (UID: \"83b71159-8ac2-4910-86ea-639368e9fe03\") " pod="openshift-infra/auto-csr-approver-29567064-8kqzc" Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.309085 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwghq\" (UniqueName: \"kubernetes.io/projected/83b71159-8ac2-4910-86ea-639368e9fe03-kube-api-access-qwghq\") pod \"auto-csr-approver-29567064-8kqzc\" (UID: \"83b71159-8ac2-4910-86ea-639368e9fe03\") " pod="openshift-infra/auto-csr-approver-29567064-8kqzc" Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.329066 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwghq\" (UniqueName: \"kubernetes.io/projected/83b71159-8ac2-4910-86ea-639368e9fe03-kube-api-access-qwghq\") pod \"auto-csr-approver-29567064-8kqzc\" (UID: \"83b71159-8ac2-4910-86ea-639368e9fe03\") " pod="openshift-infra/auto-csr-approver-29567064-8kqzc" Mar 20 16:24:00 crc kubenswrapper[4936]: I0320 16:24:00.513512 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567064-8kqzc" Mar 20 16:24:03 crc kubenswrapper[4936]: I0320 16:24:03.832641 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567064-8kqzc"] Mar 20 16:24:03 crc kubenswrapper[4936]: W0320 16:24:03.836344 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83b71159_8ac2_4910_86ea_639368e9fe03.slice/crio-12291de208b0c2bb0ad0fe07dcf446b14064f7462ad42167d1dd56c2cb199731 WatchSource:0}: Error finding container 12291de208b0c2bb0ad0fe07dcf446b14064f7462ad42167d1dd56c2cb199731: Status 404 returned error can't find the container with id 12291de208b0c2bb0ad0fe07dcf446b14064f7462ad42167d1dd56c2cb199731 Mar 20 16:24:04 crc kubenswrapper[4936]: I0320 16:24:04.111173 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" event={"ID":"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127","Type":"ContainerStarted","Data":"8aa8f7bd2be84c11c343b67ab7885d9bb71ba970c84a77f1047667af257cb1f9"} Mar 20 16:24:04 crc kubenswrapper[4936]: I0320 16:24:04.112995 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567064-8kqzc" event={"ID":"83b71159-8ac2-4910-86ea-639368e9fe03","Type":"ContainerStarted","Data":"12291de208b0c2bb0ad0fe07dcf446b14064f7462ad42167d1dd56c2cb199731"} Mar 20 16:24:04 crc kubenswrapper[4936]: I0320 16:24:04.130572 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" podStartSLOduration=1.7630627749999999 podStartE2EDuration="14.13052471s" podCreationTimestamp="2026-03-20 16:23:50 +0000 UTC" firstStartedPulling="2026-03-20 16:23:51.044159194 +0000 UTC m=+1381.990527009" lastFinishedPulling="2026-03-20 16:24:03.411621129 +0000 UTC m=+1394.357988944" observedRunningTime="2026-03-20 16:24:04.12593895 +0000 UTC m=+1395.072306765" watchObservedRunningTime="2026-03-20 16:24:04.13052471 +0000 UTC m=+1395.076892525" Mar 20 16:24:06 crc kubenswrapper[4936]: I0320 16:24:06.142506 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567064-8kqzc" event={"ID":"83b71159-8ac2-4910-86ea-639368e9fe03","Type":"ContainerStarted","Data":"f26d39a1536d4b0e64ac2186714e024431f3e1d2286b9b8bc68f44c8be5c1b3e"} Mar 20 16:24:06 crc kubenswrapper[4936]: I0320 16:24:06.165276 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567064-8kqzc" podStartSLOduration=4.68634934 podStartE2EDuration="6.165257972s" podCreationTimestamp="2026-03-20 16:24:00 +0000 UTC" firstStartedPulling="2026-03-20 16:24:03.838753669 +0000 UTC m=+1394.785121474" lastFinishedPulling="2026-03-20 16:24:05.317662291 +0000 UTC m=+1396.264030106" observedRunningTime="2026-03-20 16:24:06.164365757 +0000 UTC m=+1397.110733572" watchObservedRunningTime="2026-03-20 16:24:06.165257972 +0000 UTC m=+1397.111625787" Mar 20 16:24:07 crc kubenswrapper[4936]: I0320 16:24:07.160271 4936 generic.go:334] "Generic (PLEG): container finished" podID="83b71159-8ac2-4910-86ea-639368e9fe03" containerID="f26d39a1536d4b0e64ac2186714e024431f3e1d2286b9b8bc68f44c8be5c1b3e" exitCode=0 Mar 20 16:24:07 crc kubenswrapper[4936]: I0320 16:24:07.160631 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567064-8kqzc" event={"ID":"83b71159-8ac2-4910-86ea-639368e9fe03","Type":"ContainerDied","Data":"f26d39a1536d4b0e64ac2186714e024431f3e1d2286b9b8bc68f44c8be5c1b3e"} Mar 20 16:24:08 crc kubenswrapper[4936]: I0320 16:24:08.533812 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567064-8kqzc" Mar 20 16:24:08 crc kubenswrapper[4936]: I0320 16:24:08.595107 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwghq\" (UniqueName: \"kubernetes.io/projected/83b71159-8ac2-4910-86ea-639368e9fe03-kube-api-access-qwghq\") pod \"83b71159-8ac2-4910-86ea-639368e9fe03\" (UID: \"83b71159-8ac2-4910-86ea-639368e9fe03\") " Mar 20 16:24:08 crc kubenswrapper[4936]: I0320 16:24:08.605858 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83b71159-8ac2-4910-86ea-639368e9fe03-kube-api-access-qwghq" (OuterVolumeSpecName: "kube-api-access-qwghq") pod "83b71159-8ac2-4910-86ea-639368e9fe03" (UID: "83b71159-8ac2-4910-86ea-639368e9fe03"). InnerVolumeSpecName "kube-api-access-qwghq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:24:08 crc kubenswrapper[4936]: I0320 16:24:08.698863 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwghq\" (UniqueName: \"kubernetes.io/projected/83b71159-8ac2-4910-86ea-639368e9fe03-kube-api-access-qwghq\") on node \"crc\" DevicePath \"\"" Mar 20 16:24:09 crc kubenswrapper[4936]: I0320 16:24:09.179766 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567064-8kqzc" event={"ID":"83b71159-8ac2-4910-86ea-639368e9fe03","Type":"ContainerDied","Data":"12291de208b0c2bb0ad0fe07dcf446b14064f7462ad42167d1dd56c2cb199731"} Mar 20 16:24:09 crc kubenswrapper[4936]: I0320 16:24:09.179807 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12291de208b0c2bb0ad0fe07dcf446b14064f7462ad42167d1dd56c2cb199731" Mar 20 16:24:09 crc kubenswrapper[4936]: I0320 16:24:09.179806 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567064-8kqzc" Mar 20 16:24:09 crc kubenswrapper[4936]: I0320 16:24:09.247648 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567058-c652h"] Mar 20 16:24:09 crc kubenswrapper[4936]: I0320 16:24:09.258017 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567058-c652h"] Mar 20 16:24:09 crc kubenswrapper[4936]: I0320 16:24:09.867086 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d274515c-4ea9-459b-b934-ffd2d6982d5c" path="/var/lib/kubelet/pods/d274515c-4ea9-459b-b934-ffd2d6982d5c/volumes" Mar 20 16:24:23 crc kubenswrapper[4936]: I0320 16:24:23.320328 4936 generic.go:334] "Generic (PLEG): container finished" podID="3f1c66ea-74cb-4b08-b4fd-c1bea44ec127" containerID="8aa8f7bd2be84c11c343b67ab7885d9bb71ba970c84a77f1047667af257cb1f9" exitCode=0 Mar 20 16:24:23 crc kubenswrapper[4936]: I0320 16:24:23.320419 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" event={"ID":"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127","Type":"ContainerDied","Data":"8aa8f7bd2be84c11c343b67ab7885d9bb71ba970c84a77f1047667af257cb1f9"} Mar 20 16:24:24 crc kubenswrapper[4936]: I0320 16:24:24.451800 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:24:24 crc kubenswrapper[4936]: I0320 16:24:24.493267 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xpdzx/crc-debug-5l4z6"] Mar 20 16:24:24 crc kubenswrapper[4936]: I0320 16:24:24.494536 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xpdzx/crc-debug-5l4z6"] Mar 20 16:24:24 crc kubenswrapper[4936]: I0320 16:24:24.560898 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-host\") pod \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\" (UID: \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\") " Mar 20 16:24:24 crc kubenswrapper[4936]: I0320 16:24:24.561217 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjmtj\" (UniqueName: \"kubernetes.io/projected/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-kube-api-access-qjmtj\") pod \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\" (UID: \"3f1c66ea-74cb-4b08-b4fd-c1bea44ec127\") " Mar 20 16:24:24 crc kubenswrapper[4936]: I0320 16:24:24.561312 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-host" (OuterVolumeSpecName: "host") pod "3f1c66ea-74cb-4b08-b4fd-c1bea44ec127" (UID: "3f1c66ea-74cb-4b08-b4fd-c1bea44ec127"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:24:24 crc kubenswrapper[4936]: I0320 16:24:24.561671 4936 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-host\") on node \"crc\" DevicePath \"\"" Mar 20 16:24:24 crc kubenswrapper[4936]: I0320 16:24:24.576433 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-kube-api-access-qjmtj" (OuterVolumeSpecName: "kube-api-access-qjmtj") pod "3f1c66ea-74cb-4b08-b4fd-c1bea44ec127" (UID: "3f1c66ea-74cb-4b08-b4fd-c1bea44ec127"). InnerVolumeSpecName "kube-api-access-qjmtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:24:24 crc kubenswrapper[4936]: I0320 16:24:24.663445 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjmtj\" (UniqueName: \"kubernetes.io/projected/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127-kube-api-access-qjmtj\") on node \"crc\" DevicePath \"\"" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.340636 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fda4938a42b573233f34959f3d107e166a561ccef4cfe980b2e74e751b201a4d" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.340711 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/crc-debug-5l4z6" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.709533 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xpdzx/crc-debug-7qzwj"] Mar 20 16:24:25 crc kubenswrapper[4936]: E0320 16:24:25.710123 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83b71159-8ac2-4910-86ea-639368e9fe03" containerName="oc" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.710135 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="83b71159-8ac2-4910-86ea-639368e9fe03" containerName="oc" Mar 20 16:24:25 crc kubenswrapper[4936]: E0320 16:24:25.710173 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f1c66ea-74cb-4b08-b4fd-c1bea44ec127" containerName="container-00" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.710178 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f1c66ea-74cb-4b08-b4fd-c1bea44ec127" containerName="container-00" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.710335 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f1c66ea-74cb-4b08-b4fd-c1bea44ec127" containerName="container-00" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.710360 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="83b71159-8ac2-4910-86ea-639368e9fe03" containerName="oc" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.710946 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.717308 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xpdzx"/"default-dockercfg-85kmq" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.787969 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57290395-3b12-406e-9e98-71a9b2b4d6e8-host\") pod \"crc-debug-7qzwj\" (UID: \"57290395-3b12-406e-9e98-71a9b2b4d6e8\") " pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.788041 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hdqg\" (UniqueName: \"kubernetes.io/projected/57290395-3b12-406e-9e98-71a9b2b4d6e8-kube-api-access-9hdqg\") pod \"crc-debug-7qzwj\" (UID: \"57290395-3b12-406e-9e98-71a9b2b4d6e8\") " pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.873014 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f1c66ea-74cb-4b08-b4fd-c1bea44ec127" path="/var/lib/kubelet/pods/3f1c66ea-74cb-4b08-b4fd-c1bea44ec127/volumes" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.889935 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57290395-3b12-406e-9e98-71a9b2b4d6e8-host\") pod \"crc-debug-7qzwj\" (UID: \"57290395-3b12-406e-9e98-71a9b2b4d6e8\") " pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.890258 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hdqg\" (UniqueName: \"kubernetes.io/projected/57290395-3b12-406e-9e98-71a9b2b4d6e8-kube-api-access-9hdqg\") pod \"crc-debug-7qzwj\" (UID: \"57290395-3b12-406e-9e98-71a9b2b4d6e8\") " pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.890071 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57290395-3b12-406e-9e98-71a9b2b4d6e8-host\") pod \"crc-debug-7qzwj\" (UID: \"57290395-3b12-406e-9e98-71a9b2b4d6e8\") " pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:25 crc kubenswrapper[4936]: I0320 16:24:25.909650 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hdqg\" (UniqueName: \"kubernetes.io/projected/57290395-3b12-406e-9e98-71a9b2b4d6e8-kube-api-access-9hdqg\") pod \"crc-debug-7qzwj\" (UID: \"57290395-3b12-406e-9e98-71a9b2b4d6e8\") " pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:26 crc kubenswrapper[4936]: I0320 16:24:26.030252 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:26 crc kubenswrapper[4936]: W0320 16:24:26.085557 4936 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57290395_3b12_406e_9e98_71a9b2b4d6e8.slice/crio-44f357b7af63acad595cbba140f852e56a13a561edebffadbf6744a7c4a1d5d4 WatchSource:0}: Error finding container 44f357b7af63acad595cbba140f852e56a13a561edebffadbf6744a7c4a1d5d4: Status 404 returned error can't find the container with id 44f357b7af63acad595cbba140f852e56a13a561edebffadbf6744a7c4a1d5d4 Mar 20 16:24:26 crc kubenswrapper[4936]: I0320 16:24:26.349830 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" event={"ID":"57290395-3b12-406e-9e98-71a9b2b4d6e8","Type":"ContainerStarted","Data":"44f357b7af63acad595cbba140f852e56a13a561edebffadbf6744a7c4a1d5d4"} Mar 20 16:24:27 crc kubenswrapper[4936]: I0320 16:24:27.361948 4936 generic.go:334] "Generic (PLEG): container finished" podID="57290395-3b12-406e-9e98-71a9b2b4d6e8" containerID="cb1545f022f11ec5dcba063621baafa86e759df4b5648311ca0ab53cdb36ead9" exitCode=1 Mar 20 16:24:27 crc kubenswrapper[4936]: I0320 16:24:27.362017 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" event={"ID":"57290395-3b12-406e-9e98-71a9b2b4d6e8","Type":"ContainerDied","Data":"cb1545f022f11ec5dcba063621baafa86e759df4b5648311ca0ab53cdb36ead9"} Mar 20 16:24:27 crc kubenswrapper[4936]: I0320 16:24:27.408060 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xpdzx/crc-debug-7qzwj"] Mar 20 16:24:27 crc kubenswrapper[4936]: I0320 16:24:27.420322 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xpdzx/crc-debug-7qzwj"] Mar 20 16:24:28 crc kubenswrapper[4936]: I0320 16:24:28.478688 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:28 crc kubenswrapper[4936]: I0320 16:24:28.640870 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hdqg\" (UniqueName: \"kubernetes.io/projected/57290395-3b12-406e-9e98-71a9b2b4d6e8-kube-api-access-9hdqg\") pod \"57290395-3b12-406e-9e98-71a9b2b4d6e8\" (UID: \"57290395-3b12-406e-9e98-71a9b2b4d6e8\") " Mar 20 16:24:28 crc kubenswrapper[4936]: I0320 16:24:28.641298 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57290395-3b12-406e-9e98-71a9b2b4d6e8-host\") pod \"57290395-3b12-406e-9e98-71a9b2b4d6e8\" (UID: \"57290395-3b12-406e-9e98-71a9b2b4d6e8\") " Mar 20 16:24:28 crc kubenswrapper[4936]: I0320 16:24:28.641401 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57290395-3b12-406e-9e98-71a9b2b4d6e8-host" (OuterVolumeSpecName: "host") pod "57290395-3b12-406e-9e98-71a9b2b4d6e8" (UID: "57290395-3b12-406e-9e98-71a9b2b4d6e8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 16:24:28 crc kubenswrapper[4936]: I0320 16:24:28.642282 4936 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57290395-3b12-406e-9e98-71a9b2b4d6e8-host\") on node \"crc\" DevicePath \"\"" Mar 20 16:24:28 crc kubenswrapper[4936]: I0320 16:24:28.655195 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57290395-3b12-406e-9e98-71a9b2b4d6e8-kube-api-access-9hdqg" (OuterVolumeSpecName: "kube-api-access-9hdqg") pod "57290395-3b12-406e-9e98-71a9b2b4d6e8" (UID: "57290395-3b12-406e-9e98-71a9b2b4d6e8"). InnerVolumeSpecName "kube-api-access-9hdqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:24:28 crc kubenswrapper[4936]: I0320 16:24:28.743721 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hdqg\" (UniqueName: \"kubernetes.io/projected/57290395-3b12-406e-9e98-71a9b2b4d6e8-kube-api-access-9hdqg\") on node \"crc\" DevicePath \"\"" Mar 20 16:24:29 crc kubenswrapper[4936]: I0320 16:24:29.379928 4936 scope.go:117] "RemoveContainer" containerID="cb1545f022f11ec5dcba063621baafa86e759df4b5648311ca0ab53cdb36ead9" Mar 20 16:24:29 crc kubenswrapper[4936]: I0320 16:24:29.379961 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/crc-debug-7qzwj" Mar 20 16:24:29 crc kubenswrapper[4936]: I0320 16:24:29.864610 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57290395-3b12-406e-9e98-71a9b2b4d6e8" path="/var/lib/kubelet/pods/57290395-3b12-406e-9e98-71a9b2b4d6e8/volumes" Mar 20 16:24:40 crc kubenswrapper[4936]: I0320 16:24:40.897471 4936 scope.go:117] "RemoveContainer" containerID="0ca535be5c5f888e070a3a4225ebb14c0abc608963f4c544653e596af6be8265" Mar 20 16:24:40 crc kubenswrapper[4936]: I0320 16:24:40.954579 4936 scope.go:117] "RemoveContainer" containerID="ff195768ebd80a1cba48771ac71507daf8a3ac041103406af17fc990451d3f29" Mar 20 16:24:41 crc kubenswrapper[4936]: I0320 16:24:41.001064 4936 scope.go:117] "RemoveContainer" containerID="cf0c7f5c9b08378d74801d46ad16cbd02d8d77f2d374525945836e7538e47ffd" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.023746 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n689k"] Mar 20 16:24:52 crc kubenswrapper[4936]: E0320 16:24:52.024781 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57290395-3b12-406e-9e98-71a9b2b4d6e8" containerName="container-00" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.024798 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="57290395-3b12-406e-9e98-71a9b2b4d6e8" containerName="container-00" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.025011 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="57290395-3b12-406e-9e98-71a9b2b4d6e8" containerName="container-00" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.026919 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.038198 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n689k"] Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.210407 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-utilities\") pod \"redhat-operators-n689k\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.210470 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stvtn\" (UniqueName: \"kubernetes.io/projected/364f46cd-9f64-45ce-b42e-f95028368fad-kube-api-access-stvtn\") pod \"redhat-operators-n689k\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.210600 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-catalog-content\") pod \"redhat-operators-n689k\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.312147 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-utilities\") pod \"redhat-operators-n689k\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.312231 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stvtn\" (UniqueName: \"kubernetes.io/projected/364f46cd-9f64-45ce-b42e-f95028368fad-kube-api-access-stvtn\") pod \"redhat-operators-n689k\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.312298 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-catalog-content\") pod \"redhat-operators-n689k\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.312836 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-catalog-content\") pod \"redhat-operators-n689k\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.312855 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-utilities\") pod \"redhat-operators-n689k\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.334175 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stvtn\" (UniqueName: \"kubernetes.io/projected/364f46cd-9f64-45ce-b42e-f95028368fad-kube-api-access-stvtn\") pod \"redhat-operators-n689k\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.346509 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:24:52 crc kubenswrapper[4936]: I0320 16:24:52.849915 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n689k"] Mar 20 16:24:53 crc kubenswrapper[4936]: I0320 16:24:53.603769 4936 generic.go:334] "Generic (PLEG): container finished" podID="364f46cd-9f64-45ce-b42e-f95028368fad" containerID="af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611" exitCode=0 Mar 20 16:24:53 crc kubenswrapper[4936]: I0320 16:24:53.603823 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n689k" event={"ID":"364f46cd-9f64-45ce-b42e-f95028368fad","Type":"ContainerDied","Data":"af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611"} Mar 20 16:24:53 crc kubenswrapper[4936]: I0320 16:24:53.604009 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n689k" event={"ID":"364f46cd-9f64-45ce-b42e-f95028368fad","Type":"ContainerStarted","Data":"2ee09ce8961addf870a0ade77053042b9aa1ea7bc3796adbf63e078a977c932a"} Mar 20 16:24:54 crc kubenswrapper[4936]: I0320 16:24:54.614773 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n689k" event={"ID":"364f46cd-9f64-45ce-b42e-f95028368fad","Type":"ContainerStarted","Data":"4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52"} Mar 20 16:24:55 crc kubenswrapper[4936]: I0320 16:24:55.629298 4936 generic.go:334] "Generic (PLEG): container finished" podID="364f46cd-9f64-45ce-b42e-f95028368fad" containerID="4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52" exitCode=0 Mar 20 16:24:55 crc kubenswrapper[4936]: I0320 16:24:55.629370 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n689k" event={"ID":"364f46cd-9f64-45ce-b42e-f95028368fad","Type":"ContainerDied","Data":"4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52"} Mar 20 16:24:58 crc kubenswrapper[4936]: I0320 16:24:58.675802 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n689k" event={"ID":"364f46cd-9f64-45ce-b42e-f95028368fad","Type":"ContainerStarted","Data":"fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c"} Mar 20 16:24:58 crc kubenswrapper[4936]: I0320 16:24:58.705760 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n689k" podStartSLOduration=3.01060127 podStartE2EDuration="6.705742733s" podCreationTimestamp="2026-03-20 16:24:52 +0000 UTC" firstStartedPulling="2026-03-20 16:24:53.606167009 +0000 UTC m=+1444.552534824" lastFinishedPulling="2026-03-20 16:24:57.301308472 +0000 UTC m=+1448.247676287" observedRunningTime="2026-03-20 16:24:58.698148107 +0000 UTC m=+1449.644515952" watchObservedRunningTime="2026-03-20 16:24:58.705742733 +0000 UTC m=+1449.652110548" Mar 20 16:25:00 crc kubenswrapper[4936]: I0320 16:25:00.907559 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-7437-account-create-update-qn4hp_df4dcd90-70c5-4d83-b8a9-db6d914a8a2d/mariadb-account-create-update/0.log" Mar 20 16:25:00 crc kubenswrapper[4936]: I0320 16:25:00.919128 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-598cbc676b-tm5b2_de8a5591-dee9-46b2-81d8-8cfa2c8dce57/barbican-api/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.117576 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-598cbc676b-tm5b2_de8a5591-dee9-46b2-81d8-8cfa2c8dce57/barbican-api-log/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.264868 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-db-create-s46fs_0288b360-6ff9-4568-bc31-1351a3e55ad2/mariadb-database-create/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.351732 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-db-sync-9w2gg_61aff63e-dc1b-4854-af4a-188837ad4cf9/barbican-db-sync/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.443496 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64855595b6-t5rk7_578bb0fa-b068-4f15-9e91-b1ed317b24e6/barbican-keystone-listener/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.513669 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64855595b6-t5rk7_578bb0fa-b068-4f15-9e91-b1ed317b24e6/barbican-keystone-listener-log/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.658153 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7df458d599-sn4pk_46c60de1-d091-43da-9f0e-9fa0e31bb556/barbican-worker/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.715252 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7df458d599-sn4pk_46c60de1-d091-43da-9f0e-9fa0e31bb556/barbican-worker-log/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.891663 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e03d33f8-2aec-41af-8c22-48b41893d7b9/ceilometer-central-agent/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.930472 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e03d33f8-2aec-41af-8c22-48b41893d7b9/ceilometer-notification-agent/0.log" Mar 20 16:25:01 crc kubenswrapper[4936]: I0320 16:25:01.970175 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e03d33f8-2aec-41af-8c22-48b41893d7b9/proxy-httpd/0.log" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.046318 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e03d33f8-2aec-41af-8c22-48b41893d7b9/sg-core/0.log" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.179577 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-63a4-account-create-update-wng8g_df8f8d6e-b8f0-4148-940e-712399f97ab1/mariadb-account-create-update/0.log" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.294222 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_2314f249-26ea-45c5-8225-ca99d8a4f82e/cinder-api/0.log" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.347152 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.348245 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.366891 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_2314f249-26ea-45c5-8225-ca99d8a4f82e/cinder-api-log/0.log" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.509410 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-db-create-p249h_2726b8ff-699c-419e-ae52-0868c355ea23/mariadb-database-create/0.log" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.604109 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-db-sync-n7kq7_047fb63e-64a8-4280-9846-659575038df3/cinder-db-sync/0.log" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.721688 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_46cbf2a6-9c75-49a6-b7b1-31fcf600d874/cinder-scheduler/0.log" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.860282 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_46cbf2a6-9c75-49a6-b7b1-31fcf600d874/probe/0.log" Mar 20 16:25:02 crc kubenswrapper[4936]: I0320 16:25:02.920259 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-89c5cd4d5-ln75g_3eed9143-556a-4df3-bef4-1eb97bf9ffe4/init/0.log" Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.103805 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-89c5cd4d5-ln75g_3eed9143-556a-4df3-bef4-1eb97bf9ffe4/init/0.log" Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.137539 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-89c5cd4d5-ln75g_3eed9143-556a-4df3-bef4-1eb97bf9ffe4/dnsmasq-dns/0.log" Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.155690 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-05f5-account-create-update-9vd8r_11a87bf1-c0b8-47a2-8ccb-faf322bebfb1/mariadb-account-create-update/0.log" Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.305669 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-db-create-9vpbv_812b9782-84b1-41fc-939e-30e82c1b5e13/mariadb-database-create/0.log" Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.384843 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-db-sync-sg9df_a817ffe2-32a4-4c37-9342-5d642d179ea7/glance-db-sync/0.log" Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.408694 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n689k" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="registry-server" probeResult="failure" output=< Mar 20 16:25:03 crc kubenswrapper[4936]: timeout: failed to connect service ":50051" within 1s Mar 20 16:25:03 crc kubenswrapper[4936]: > Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.605677 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc2d8076-f00b-4d31-89ba-4f15a9d6d50a/glance-log/0.log" Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.610844 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc2d8076-f00b-4d31-89ba-4f15a9d6d50a/glance-httpd/0.log" Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.768731 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3640384f-37b8-43b3-8fe2-10359b9482a8/glance-httpd/0.log" Mar 20 16:25:03 crc kubenswrapper[4936]: I0320 16:25:03.835442 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3640384f-37b8-43b3-8fe2-10359b9482a8/glance-log/0.log" Mar 20 16:25:04 crc kubenswrapper[4936]: I0320 16:25:04.000984 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6ff4f57d48-qp24t_7f1bd9ee-e0b7-4ea9-b3ca-97987941f149/horizon/0.log" Mar 20 16:25:04 crc kubenswrapper[4936]: I0320 16:25:04.015207 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6ff4f57d48-qp24t_7f1bd9ee-e0b7-4ea9-b3ca-97987941f149/horizon-log/0.log" Mar 20 16:25:04 crc kubenswrapper[4936]: I0320 16:25:04.174930 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-768674d4fd-4vgnw_9c24263c-e99e-4126-9428-55ea6d8cc908/keystone-api/0.log" Mar 20 16:25:04 crc kubenswrapper[4936]: I0320 16:25:04.254086 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7948-account-create-update-8trc2_a0aa18c2-d52b-4671-996b-985c848a748f/mariadb-account-create-update/0.log" Mar 20 16:25:04 crc kubenswrapper[4936]: I0320 16:25:04.403884 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-bootstrap-w6hwd_3cd7ee68-8aa9-4b2c-9d06-c4de9698b755/keystone-bootstrap/0.log" Mar 20 16:25:04 crc kubenswrapper[4936]: I0320 16:25:04.498943 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-db-create-zphvf_434fff7b-3810-45b8-8028-98a9d77d3f15/mariadb-database-create/0.log" Mar 20 16:25:04 crc kubenswrapper[4936]: I0320 16:25:04.628874 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-db-sync-wfxxc_e13cc3e4-9442-40ec-b43f-cb968e26a0e1/keystone-db-sync/0.log" Mar 20 16:25:04 crc kubenswrapper[4936]: I0320 16:25:04.781583 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_4cc720ee-42c3-4ce6-87de-045d881d730d/kube-state-metrics/0.log" Mar 20 16:25:05 crc kubenswrapper[4936]: I0320 16:25:05.065895 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-698b9689fc-cph26_7c402d19-99d9-49b2-9304-12ef96435d7c/neutron-api/0.log" Mar 20 16:25:05 crc kubenswrapper[4936]: I0320 16:25:05.069379 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-698b9689fc-cph26_7c402d19-99d9-49b2-9304-12ef96435d7c/neutron-httpd/0.log" Mar 20 16:25:05 crc kubenswrapper[4936]: I0320 16:25:05.302384 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-ab53-account-create-update-rrk7z_b802c0e2-15be-4604-b64c-3b7441a1096d/mariadb-account-create-update/0.log" Mar 20 16:25:05 crc kubenswrapper[4936]: I0320 16:25:05.478279 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-db-create-m5pnp_da80eafb-7c87-4b87-89ae-beee42d9b256/mariadb-database-create/0.log" Mar 20 16:25:05 crc kubenswrapper[4936]: I0320 16:25:05.578073 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-db-sync-4t65w_5421152c-d474-428f-9554-96e036719655/neutron-db-sync/0.log" Mar 20 16:25:05 crc kubenswrapper[4936]: I0320 16:25:05.788117 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_02a89fea-4ef7-41e3-a914-5cec1d6c7439/nova-api-log/0.log" Mar 20 16:25:05 crc kubenswrapper[4936]: I0320 16:25:05.853304 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_02a89fea-4ef7-41e3-a914-5cec1d6c7439/nova-api-api/0.log" Mar 20 16:25:05 crc kubenswrapper[4936]: I0320 16:25:05.888922 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-06af-account-create-update-t8zxx_d9093776-e539-4e6e-ac2e-01da662cbe84/mariadb-account-create-update/0.log" Mar 20 16:25:06 crc kubenswrapper[4936]: I0320 16:25:06.190209 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-685b-account-create-update-j7zd9_f978ca59-3511-4a0f-9ae4-944e843f1c45/mariadb-account-create-update/0.log" Mar 20 16:25:06 crc kubenswrapper[4936]: I0320 16:25:06.212078 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-db-create-ktpms_04a90ca0-4946-4286-86b1-ba337895f3e8/mariadb-database-create/0.log" Mar 20 16:25:06 crc kubenswrapper[4936]: I0320 16:25:06.446689 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-cell-mapping-k54pt_446d671c-a5e1-4430-9aa6-262bb3a9b2d0/nova-manage/0.log" Mar 20 16:25:06 crc kubenswrapper[4936]: I0320 16:25:06.570969 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d4f2d11d-47b6-4791-88c9-25911a5b6bc7/nova-cell0-conductor-conductor/0.log" Mar 20 16:25:06 crc kubenswrapper[4936]: I0320 16:25:06.695062 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-db-sync-hspgq_49bf6254-7f19-41f0-969b-fc8b338ece00/nova-cell0-conductor-db-sync/0.log" Mar 20 16:25:06 crc kubenswrapper[4936]: I0320 16:25:06.800356 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-db-create-f9qdx_ee8f6e3e-3bed-4aca-b7b4-98611a317a19/mariadb-database-create/0.log" Mar 20 16:25:07 crc kubenswrapper[4936]: I0320 16:25:07.015092 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-416f-account-create-update-dk82t_a4111fca-7b2f-456e-9591-e9f4b8f764a2/mariadb-account-create-update/0.log" Mar 20 16:25:07 crc kubenswrapper[4936]: I0320 16:25:07.096998 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-cell-mapping-snhvn_3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c/nova-manage/0.log" Mar 20 16:25:07 crc kubenswrapper[4936]: I0320 16:25:07.317947 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7810d8f5-e6ff-402c-a916-42e6125f2d6e/nova-cell1-conductor-conductor/0.log" Mar 20 16:25:07 crc kubenswrapper[4936]: I0320 16:25:07.357646 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-db-sync-fvb75_6071aca2-12af-427f-88a0-5661b147f6d8/nova-cell1-conductor-db-sync/0.log" Mar 20 16:25:07 crc kubenswrapper[4936]: I0320 16:25:07.576690 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-db-create-tv2ph_ec06ab44-bb2e-4add-8408-49d78727c812/mariadb-database-create/0.log" Mar 20 16:25:07 crc kubenswrapper[4936]: I0320 16:25:07.645114 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_556eac71-c652-4652-8b3e-2b208ca34704/nova-cell1-novncproxy-novncproxy/0.log" Mar 20 16:25:07 crc kubenswrapper[4936]: I0320 16:25:07.991949 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7e89b3de-e3f1-4258-9f7d-937b8159f213/nova-metadata-log/0.log" Mar 20 16:25:08 crc kubenswrapper[4936]: I0320 16:25:08.013222 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7e89b3de-e3f1-4258-9f7d-937b8159f213/nova-metadata-metadata/0.log" Mar 20 16:25:08 crc kubenswrapper[4936]: I0320 16:25:08.230033 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_82bd48e0-4418-4657-9082-11140724699f/mysql-bootstrap/0.log" Mar 20 16:25:08 crc kubenswrapper[4936]: I0320 16:25:08.268783 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e3bf3212-25ad-4ec7-bb5d-c162c5428220/nova-scheduler-scheduler/0.log" Mar 20 16:25:08 crc kubenswrapper[4936]: I0320 16:25:08.420117 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_82bd48e0-4418-4657-9082-11140724699f/mysql-bootstrap/0.log" Mar 20 16:25:08 crc kubenswrapper[4936]: I0320 16:25:08.486840 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_82bd48e0-4418-4657-9082-11140724699f/galera/0.log" Mar 20 16:25:08 crc kubenswrapper[4936]: I0320 16:25:08.542101 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2d8d2157-625a-4b47-a722-1ccbe9236ef6/mysql-bootstrap/0.log" Mar 20 16:25:08 crc kubenswrapper[4936]: I0320 16:25:08.733873 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2d8d2157-625a-4b47-a722-1ccbe9236ef6/mysql-bootstrap/0.log" Mar 20 16:25:08 crc kubenswrapper[4936]: I0320 16:25:08.739923 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2d8d2157-625a-4b47-a722-1ccbe9236ef6/galera/0.log" Mar 20 16:25:08 crc kubenswrapper[4936]: I0320 16:25:08.787523 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_961ea539-f955-46fd-8afa-8ec5b42133eb/openstackclient/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.015751 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vpzks_d6b58590-f075-4e5c-8bad-6f8ebb368c93/openstack-network-exporter/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.104177 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tp2gk_a660d73e-dab8-4751-8219-c91f6017b574/ovsdb-server-init/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.330678 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tp2gk_a660d73e-dab8-4751-8219-c91f6017b574/ovsdb-server/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.395320 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tp2gk_a660d73e-dab8-4751-8219-c91f6017b574/ovsdb-server-init/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.400602 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tp2gk_a660d73e-dab8-4751-8219-c91f6017b574/ovs-vswitchd/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.538947 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-rnftv_cd2037c4-bb05-4f92-9a4a-f71c5da11ebd/ovn-controller/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.632947 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7f4f09dd-f403-4f3b-ac44-129df52fd796/ovn-northd/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.686632 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7f4f09dd-f403-4f3b-ac44-129df52fd796/openstack-network-exporter/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.874011 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7/openstack-network-exporter/0.log" Mar 20 16:25:09 crc kubenswrapper[4936]: I0320 16:25:09.909958 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1ba28ba5-07dc-4f0c-ab8a-2d23b6e143e7/ovsdbserver-nb/0.log" Mar 20 16:25:10 crc kubenswrapper[4936]: I0320 16:25:10.074406 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7b0e8ad7-4143-431d-a022-c467ee2d09ac/ovsdbserver-sb/0.log" Mar 20 16:25:10 crc kubenswrapper[4936]: I0320 16:25:10.095821 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7b0e8ad7-4143-431d-a022-c467ee2d09ac/openstack-network-exporter/0.log" Mar 20 16:25:10 crc kubenswrapper[4936]: I0320 16:25:10.246613 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-584c7d8894-w4d85_1405ee40-3191-4a1e-9ad7-27b12cd5abaf/placement-api/0.log" Mar 20 16:25:10 crc kubenswrapper[4936]: I0320 16:25:10.322158 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-584c7d8894-w4d85_1405ee40-3191-4a1e-9ad7-27b12cd5abaf/placement-log/0.log" Mar 20 16:25:10 crc kubenswrapper[4936]: I0320 16:25:10.453348 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-962c-account-create-update-nbrwz_c48f3728-76d9-43c5-8053-b27ac3751ecb/mariadb-account-create-update/0.log" Mar 20 16:25:10 crc kubenswrapper[4936]: I0320 16:25:10.563134 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-db-create-jvq45_b8df2114-9d0b-43c8-91cd-a60355f7238b/mariadb-database-create/0.log" Mar 20 16:25:10 crc kubenswrapper[4936]: I0320 16:25:10.790638 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-db-sync-j4d5s_b87c4418-b13e-41e2-bbd8-d7dfb057a594/placement-db-sync/0.log" Mar 20 16:25:10 crc kubenswrapper[4936]: I0320 16:25:10.835083 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_18cfd431-66b8-49d8-8dd9-b2184415c006/setup-container/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.004858 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_18cfd431-66b8-49d8-8dd9-b2184415c006/setup-container/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.032284 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_18cfd431-66b8-49d8-8dd9-b2184415c006/rabbitmq/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.126570 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3b3a9af8-80db-4edc-9f6a-42b641c25d71/setup-container/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.263630 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3b3a9af8-80db-4edc-9f6a-42b641c25d71/setup-container/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.341578 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3b3a9af8-80db-4edc-9f6a-42b641c25d71/rabbitmq/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.363480 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_root-account-create-update-cjk9m_3aece10e-0797-48c2-aac7-7a4b12cf0cf5/mariadb-account-create-update/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.557144 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-74b5cc8579-wqhwv_9284dab5-bdd1-4792-8944-0f9c820a927d/proxy-httpd/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.567181 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-74b5cc8579-wqhwv_9284dab5-bdd1-4792-8944-0f9c820a927d/proxy-server/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.713903 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6k8ph_070d6439-9bd3-4fd8-b61a-df1f03badb79/swift-ring-rebalance/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.775767 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/account-auditor/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.902493 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/account-reaper/0.log" Mar 20 16:25:11 crc kubenswrapper[4936]: I0320 16:25:11.952940 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/account-replicator/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.010403 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/account-server/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.036862 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/container-auditor/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.113006 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/container-replicator/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.236527 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/container-server/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.239394 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/container-updater/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.335955 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/object-auditor/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.340108 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/object-expirer/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.433918 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/object-replicator/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.444617 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/object-server/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.527287 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/object-updater/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.607169 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/rsync/0.log" Mar 20 16:25:12 crc kubenswrapper[4936]: I0320 16:25:12.626955 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b1ea233b-41e7-49ac-a878-aa038b350252/swift-recon-cron/0.log" Mar 20 16:25:13 crc kubenswrapper[4936]: I0320 16:25:13.400118 4936 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n689k" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="registry-server" probeResult="failure" output=< Mar 20 16:25:13 crc kubenswrapper[4936]: timeout: failed to connect service ":50051" within 1s Mar 20 16:25:13 crc kubenswrapper[4936]: > Mar 20 16:25:13 crc kubenswrapper[4936]: I0320 16:25:13.840273 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6482c360-33e0-4e79-9728-6da33e8adbd1/memcached/0.log" Mar 20 16:25:22 crc kubenswrapper[4936]: I0320 16:25:22.409874 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:25:22 crc kubenswrapper[4936]: I0320 16:25:22.464923 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.025236 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n689k"] Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.025777 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n689k" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="registry-server" containerID="cri-o://fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c" gracePeriod=2 Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.478460 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.534936 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-utilities\") pod \"364f46cd-9f64-45ce-b42e-f95028368fad\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.535429 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-catalog-content\") pod \"364f46cd-9f64-45ce-b42e-f95028368fad\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.535523 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stvtn\" (UniqueName: \"kubernetes.io/projected/364f46cd-9f64-45ce-b42e-f95028368fad-kube-api-access-stvtn\") pod \"364f46cd-9f64-45ce-b42e-f95028368fad\" (UID: \"364f46cd-9f64-45ce-b42e-f95028368fad\") " Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.535768 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-utilities" (OuterVolumeSpecName: "utilities") pod "364f46cd-9f64-45ce-b42e-f95028368fad" (UID: "364f46cd-9f64-45ce-b42e-f95028368fad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.536071 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.541712 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/364f46cd-9f64-45ce-b42e-f95028368fad-kube-api-access-stvtn" (OuterVolumeSpecName: "kube-api-access-stvtn") pod "364f46cd-9f64-45ce-b42e-f95028368fad" (UID: "364f46cd-9f64-45ce-b42e-f95028368fad"). InnerVolumeSpecName "kube-api-access-stvtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.639091 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stvtn\" (UniqueName: \"kubernetes.io/projected/364f46cd-9f64-45ce-b42e-f95028368fad-kube-api-access-stvtn\") on node \"crc\" DevicePath \"\"" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.688685 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "364f46cd-9f64-45ce-b42e-f95028368fad" (UID: "364f46cd-9f64-45ce-b42e-f95028368fad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.742286 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/364f46cd-9f64-45ce-b42e-f95028368fad-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.909284 4936 generic.go:334] "Generic (PLEG): container finished" podID="364f46cd-9f64-45ce-b42e-f95028368fad" containerID="fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c" exitCode=0 Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.909331 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n689k" event={"ID":"364f46cd-9f64-45ce-b42e-f95028368fad","Type":"ContainerDied","Data":"fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c"} Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.909362 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n689k" event={"ID":"364f46cd-9f64-45ce-b42e-f95028368fad","Type":"ContainerDied","Data":"2ee09ce8961addf870a0ade77053042b9aa1ea7bc3796adbf63e078a977c932a"} Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.909370 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n689k" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.909382 4936 scope.go:117] "RemoveContainer" containerID="fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.954455 4936 scope.go:117] "RemoveContainer" containerID="4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52" Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.959234 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n689k"] Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.975077 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n689k"] Mar 20 16:25:24 crc kubenswrapper[4936]: I0320 16:25:24.986182 4936 scope.go:117] "RemoveContainer" containerID="af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611" Mar 20 16:25:25 crc kubenswrapper[4936]: I0320 16:25:25.028532 4936 scope.go:117] "RemoveContainer" containerID="fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c" Mar 20 16:25:25 crc kubenswrapper[4936]: E0320 16:25:25.028964 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c\": container with ID starting with fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c not found: ID does not exist" containerID="fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c" Mar 20 16:25:25 crc kubenswrapper[4936]: I0320 16:25:25.029006 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c"} err="failed to get container status \"fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c\": rpc error: code = NotFound desc = could not find container \"fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c\": container with ID starting with fcd51482b822007068350e6355e86bb78fe26304ff4c0ff18d2b3f8cca2afc6c not found: ID does not exist" Mar 20 16:25:25 crc kubenswrapper[4936]: I0320 16:25:25.029031 4936 scope.go:117] "RemoveContainer" containerID="4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52" Mar 20 16:25:25 crc kubenswrapper[4936]: E0320 16:25:25.029424 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52\": container with ID starting with 4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52 not found: ID does not exist" containerID="4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52" Mar 20 16:25:25 crc kubenswrapper[4936]: I0320 16:25:25.029452 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52"} err="failed to get container status \"4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52\": rpc error: code = NotFound desc = could not find container \"4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52\": container with ID starting with 4e63bdf61ff1b041b9600b4a80b9e068165077775ee928425c880a83254eaa52 not found: ID does not exist" Mar 20 16:25:25 crc kubenswrapper[4936]: I0320 16:25:25.029476 4936 scope.go:117] "RemoveContainer" containerID="af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611" Mar 20 16:25:25 crc kubenswrapper[4936]: E0320 16:25:25.029841 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611\": container with ID starting with af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611 not found: ID does not exist" containerID="af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611" Mar 20 16:25:25 crc kubenswrapper[4936]: I0320 16:25:25.029872 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611"} err="failed to get container status \"af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611\": rpc error: code = NotFound desc = could not find container \"af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611\": container with ID starting with af85e77038aed992210ba08359973dc6f30976cea09e8fce48ef09823fd9a611 not found: ID does not exist" Mar 20 16:25:25 crc kubenswrapper[4936]: I0320 16:25:25.871915 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" path="/var/lib/kubelet/pods/364f46cd-9f64-45ce-b42e-f95028368fad/volumes" Mar 20 16:25:34 crc kubenswrapper[4936]: I0320 16:25:34.809712 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz_cada20db-602e-45c5-be95-65c083299de3/util/0.log" Mar 20 16:25:34 crc kubenswrapper[4936]: I0320 16:25:34.920323 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz_cada20db-602e-45c5-be95-65c083299de3/util/0.log" Mar 20 16:25:34 crc kubenswrapper[4936]: I0320 16:25:34.968160 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz_cada20db-602e-45c5-be95-65c083299de3/pull/0.log" Mar 20 16:25:34 crc kubenswrapper[4936]: I0320 16:25:34.996072 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz_cada20db-602e-45c5-be95-65c083299de3/pull/0.log" Mar 20 16:25:35 crc kubenswrapper[4936]: I0320 16:25:35.157475 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz_cada20db-602e-45c5-be95-65c083299de3/pull/0.log" Mar 20 16:25:35 crc kubenswrapper[4936]: I0320 16:25:35.189852 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz_cada20db-602e-45c5-be95-65c083299de3/extract/0.log" Mar 20 16:25:35 crc kubenswrapper[4936]: I0320 16:25:35.224006 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d0d95ff7b764438735c5134025d337ce5d025c28ea0a327da9967842bnkgtz_cada20db-602e-45c5-be95-65c083299de3/util/0.log" Mar 20 16:25:35 crc kubenswrapper[4936]: I0320 16:25:35.457082 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-4pkh7_4cb28ddc-4011-42fb-b8c9-05a2e358330a/manager/0.log" Mar 20 16:25:35 crc kubenswrapper[4936]: I0320 16:25:35.609597 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-qk46l_aeb73df4-0e66-4108-b4a2-9aa41de2fcd0/manager/0.log" Mar 20 16:25:35 crc kubenswrapper[4936]: I0320 16:25:35.757334 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-t7zlc_516b1d4c-72c7-4af0-81bf-489c2a598f8a/manager/0.log" Mar 20 16:25:36 crc kubenswrapper[4936]: I0320 16:25:36.109653 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-fxfnf_f7dbf249-277f-4cb7-8148-7b6062cc5414/manager/0.log" Mar 20 16:25:36 crc kubenswrapper[4936]: I0320 16:25:36.258635 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-674vf_3f99cb10-909b-4f34-9b66-95ce7440af5d/manager/0.log" Mar 20 16:25:36 crc kubenswrapper[4936]: I0320 16:25:36.277496 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-t7n5g_d8898bff-2619-4d34-8998-36ef5aef1c5b/manager/0.log" Mar 20 16:25:36 crc kubenswrapper[4936]: I0320 16:25:36.511952 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-c7cqh_8f7f9e3c-7ad9-409e-bc82-6c74e3101630/manager/0.log" Mar 20 16:25:36 crc kubenswrapper[4936]: I0320 16:25:36.584046 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c7774cdcf-p6qdv_91412ce4-2ae2-4d80-b2d9-1f9825eaeb35/manager/0.log" Mar 20 16:25:36 crc kubenswrapper[4936]: I0320 16:25:36.716509 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-5mq8g_0fbd2515-7268-4434-b002-f8e44a0e7b0d/manager/0.log" Mar 20 16:25:36 crc kubenswrapper[4936]: I0320 16:25:36.722404 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-jdd86_9f92f032-d633-4845-8d38-5cbe5bd39313/manager/0.log" Mar 20 16:25:36 crc kubenswrapper[4936]: I0320 16:25:36.943489 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-hrgvc_80f063a2-e5ad-462b-9a61-06eadd11d477/manager/0.log" Mar 20 16:25:36 crc kubenswrapper[4936]: I0320 16:25:36.982152 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-f48h2_7aa0f202-e676-4f55-a20e-64ffc6b90339/manager/0.log" Mar 20 16:25:37 crc kubenswrapper[4936]: I0320 16:25:37.186005 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-cpjvf_e9f3d1f0-085a-42b4-92c2-b69e78f69667/manager/0.log" Mar 20 16:25:37 crc kubenswrapper[4936]: I0320 16:25:37.206122 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-dzfhp_d10fa5e0-be5a-4b79-85d9-a7451ec86255/manager/0.log" Mar 20 16:25:37 crc kubenswrapper[4936]: I0320 16:25:37.363935 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-86657c54f5gkwd5_a92eb460-57c4-49c7-bd39-2ddb71062d32/manager/0.log" Mar 20 16:25:37 crc kubenswrapper[4936]: I0320 16:25:37.471515 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-658f6bdff6-vrsv9_c4efa574-8274-4c0b-9463-d076e31d2a7e/operator/0.log" Mar 20 16:25:37 crc kubenswrapper[4936]: I0320 16:25:37.681030 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-c4nw2_0401145a-6ad6-44e3-b2ab-ab333cc2c118/registry-server/0.log" Mar 20 16:25:37 crc kubenswrapper[4936]: I0320 16:25:37.950590 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-clwss_ae8fec9a-8dd7-481d-a053-9185bdd5aa46/manager/0.log" Mar 20 16:25:38 crc kubenswrapper[4936]: I0320 16:25:38.062434 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-9n5lb_830190cb-826c-4da0-a31c-f845de9707a9/manager/0.log" Mar 20 16:25:38 crc kubenswrapper[4936]: I0320 16:25:38.217698 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-fvbgv_34327669-3184-41ae-b81f-60b97f97a74c/manager/0.log" Mar 20 16:25:38 crc kubenswrapper[4936]: I0320 16:25:38.339182 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-889c6bbdb-678j6_79279548-dd43-4465-8d51-399bf8ed16ab/manager/0.log" Mar 20 16:25:38 crc kubenswrapper[4936]: I0320 16:25:38.411307 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-5pfk5_7503f97b-163f-40bd-bc93-7d8310c51965/manager/0.log" Mar 20 16:25:38 crc kubenswrapper[4936]: I0320 16:25:38.537509 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-nbhxk_1420e79c-a73c-4ed8-902e-eb656a065140/manager/0.log" Mar 20 16:25:38 crc kubenswrapper[4936]: I0320 16:25:38.602275 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-q7pkz_6944a43f-5df3-47b3-ac8c-a6f90eb0cfc7/manager/0.log" Mar 20 16:25:41 crc kubenswrapper[4936]: I0320 16:25:41.122452 4936 scope.go:117] "RemoveContainer" containerID="500981b2dbc9c7c4e6e72392c7242fa937724f289025c56305f488d06c177e0c" Mar 20 16:25:57 crc kubenswrapper[4936]: I0320 16:25:57.620715 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-vm2rn_cd07f1f3-ea79-4612-af59-91151c9bb06f/control-plane-machine-set-operator/0.log" Mar 20 16:25:57 crc kubenswrapper[4936]: I0320 16:25:57.821863 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ssr5n_d815687f-6e2a-44c8-bf71-c9718db7ba67/kube-rbac-proxy/0.log" Mar 20 16:25:57 crc kubenswrapper[4936]: I0320 16:25:57.903339 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ssr5n_d815687f-6e2a-44c8-bf71-c9718db7ba67/machine-api-operator/0.log" Mar 20 16:25:58 crc kubenswrapper[4936]: I0320 16:25:58.855222 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:25:58 crc kubenswrapper[4936]: I0320 16:25:58.855301 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.140662 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567066-d4pbg"] Mar 20 16:26:00 crc kubenswrapper[4936]: E0320 16:26:00.141079 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="extract-content" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.141092 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="extract-content" Mar 20 16:26:00 crc kubenswrapper[4936]: E0320 16:26:00.141118 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="extract-utilities" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.141124 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="extract-utilities" Mar 20 16:26:00 crc kubenswrapper[4936]: E0320 16:26:00.141145 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="registry-server" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.141153 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="registry-server" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.141329 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="364f46cd-9f64-45ce-b42e-f95028368fad" containerName="registry-server" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.142006 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567066-d4pbg" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.143905 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.144183 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.144431 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.148638 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567066-d4pbg"] Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.225121 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgdfw\" (UniqueName: \"kubernetes.io/projected/70fe6af5-2344-49de-a870-913d9edffc8b-kube-api-access-fgdfw\") pod \"auto-csr-approver-29567066-d4pbg\" (UID: \"70fe6af5-2344-49de-a870-913d9edffc8b\") " pod="openshift-infra/auto-csr-approver-29567066-d4pbg" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.326547 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgdfw\" (UniqueName: \"kubernetes.io/projected/70fe6af5-2344-49de-a870-913d9edffc8b-kube-api-access-fgdfw\") pod \"auto-csr-approver-29567066-d4pbg\" (UID: \"70fe6af5-2344-49de-a870-913d9edffc8b\") " pod="openshift-infra/auto-csr-approver-29567066-d4pbg" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.346827 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgdfw\" (UniqueName: \"kubernetes.io/projected/70fe6af5-2344-49de-a870-913d9edffc8b-kube-api-access-fgdfw\") pod \"auto-csr-approver-29567066-d4pbg\" (UID: \"70fe6af5-2344-49de-a870-913d9edffc8b\") " pod="openshift-infra/auto-csr-approver-29567066-d4pbg" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.514480 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567066-d4pbg" Mar 20 16:26:00 crc kubenswrapper[4936]: I0320 16:26:00.976655 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567066-d4pbg"] Mar 20 16:26:01 crc kubenswrapper[4936]: I0320 16:26:01.220252 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567066-d4pbg" event={"ID":"70fe6af5-2344-49de-a870-913d9edffc8b","Type":"ContainerStarted","Data":"a5998c9e5901d34fe8dd4ed219ac3a2c910359ef48ebf8713a85a69ffb4207a3"} Mar 20 16:26:03 crc kubenswrapper[4936]: I0320 16:26:03.237036 4936 generic.go:334] "Generic (PLEG): container finished" podID="70fe6af5-2344-49de-a870-913d9edffc8b" containerID="38f4cd7ab4f178d8e5ada3d9e6533d3eda21750895f91e2443e3de6b0efd684a" exitCode=0 Mar 20 16:26:03 crc kubenswrapper[4936]: I0320 16:26:03.237144 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567066-d4pbg" event={"ID":"70fe6af5-2344-49de-a870-913d9edffc8b","Type":"ContainerDied","Data":"38f4cd7ab4f178d8e5ada3d9e6533d3eda21750895f91e2443e3de6b0efd684a"} Mar 20 16:26:04 crc kubenswrapper[4936]: I0320 16:26:04.660058 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567066-d4pbg" Mar 20 16:26:04 crc kubenswrapper[4936]: I0320 16:26:04.805190 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgdfw\" (UniqueName: \"kubernetes.io/projected/70fe6af5-2344-49de-a870-913d9edffc8b-kube-api-access-fgdfw\") pod \"70fe6af5-2344-49de-a870-913d9edffc8b\" (UID: \"70fe6af5-2344-49de-a870-913d9edffc8b\") " Mar 20 16:26:04 crc kubenswrapper[4936]: I0320 16:26:04.811220 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70fe6af5-2344-49de-a870-913d9edffc8b-kube-api-access-fgdfw" (OuterVolumeSpecName: "kube-api-access-fgdfw") pod "70fe6af5-2344-49de-a870-913d9edffc8b" (UID: "70fe6af5-2344-49de-a870-913d9edffc8b"). InnerVolumeSpecName "kube-api-access-fgdfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:26:04 crc kubenswrapper[4936]: I0320 16:26:04.907292 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgdfw\" (UniqueName: \"kubernetes.io/projected/70fe6af5-2344-49de-a870-913d9edffc8b-kube-api-access-fgdfw\") on node \"crc\" DevicePath \"\"" Mar 20 16:26:05 crc kubenswrapper[4936]: I0320 16:26:05.256783 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567066-d4pbg" event={"ID":"70fe6af5-2344-49de-a870-913d9edffc8b","Type":"ContainerDied","Data":"a5998c9e5901d34fe8dd4ed219ac3a2c910359ef48ebf8713a85a69ffb4207a3"} Mar 20 16:26:05 crc kubenswrapper[4936]: I0320 16:26:05.257141 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5998c9e5901d34fe8dd4ed219ac3a2c910359ef48ebf8713a85a69ffb4207a3" Mar 20 16:26:05 crc kubenswrapper[4936]: I0320 16:26:05.256844 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567066-d4pbg" Mar 20 16:26:05 crc kubenswrapper[4936]: I0320 16:26:05.726793 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567060-gbvj5"] Mar 20 16:26:05 crc kubenswrapper[4936]: I0320 16:26:05.734448 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567060-gbvj5"] Mar 20 16:26:05 crc kubenswrapper[4936]: I0320 16:26:05.864818 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05b79d13-f57b-455c-b5fe-b469c9bc0f45" path="/var/lib/kubelet/pods/05b79d13-f57b-455c-b5fe-b469c9bc0f45/volumes" Mar 20 16:26:09 crc kubenswrapper[4936]: I0320 16:26:09.840798 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-lm2fv_084aa53e-7882-49d2-afac-d70afccc98b8/cert-manager-controller/0.log" Mar 20 16:26:10 crc kubenswrapper[4936]: I0320 16:26:10.017969 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-kcmnv_0f954785-a2e7-46c3-9958-d243fa27da2c/cert-manager-cainjector/0.log" Mar 20 16:26:10 crc kubenswrapper[4936]: I0320 16:26:10.042037 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-z9dsw_b04f08ab-f31c-4229-8158-a589251be69c/cert-manager-webhook/0.log" Mar 20 16:26:21 crc kubenswrapper[4936]: I0320 16:26:21.740661 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-dk6hh_00920150-af2f-46d4-a3cf-bdfea2640c89/nmstate-console-plugin/0.log" Mar 20 16:26:21 crc kubenswrapper[4936]: I0320 16:26:21.911685 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-h59z7_36c16ea1-7b8b-4175-8647-92fc61ee66d5/nmstate-handler/0.log" Mar 20 16:26:21 crc kubenswrapper[4936]: I0320 16:26:21.984968 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-mmtw4_f0f1b197-b717-44db-aed0-98ae53aba831/kube-rbac-proxy/0.log" Mar 20 16:26:22 crc kubenswrapper[4936]: I0320 16:26:22.036157 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-mmtw4_f0f1b197-b717-44db-aed0-98ae53aba831/nmstate-metrics/0.log" Mar 20 16:26:22 crc kubenswrapper[4936]: I0320 16:26:22.171916 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-b94r2_e22309a2-eb8b-4a9e-8f56-43c169803a33/nmstate-operator/0.log" Mar 20 16:26:22 crc kubenswrapper[4936]: I0320 16:26:22.233184 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-dqbtr_8b4af217-1e38-47e1-b306-a9d4d7ed90f8/nmstate-webhook/0.log" Mar 20 16:26:28 crc kubenswrapper[4936]: I0320 16:26:28.854696 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:26:28 crc kubenswrapper[4936]: I0320 16:26:28.855202 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.304526 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ppdjl"] Mar 20 16:26:34 crc kubenswrapper[4936]: E0320 16:26:34.305666 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70fe6af5-2344-49de-a870-913d9edffc8b" containerName="oc" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.305681 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="70fe6af5-2344-49de-a870-913d9edffc8b" containerName="oc" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.305914 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="70fe6af5-2344-49de-a870-913d9edffc8b" containerName="oc" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.307418 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.326850 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppdjl"] Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.376071 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-catalog-content\") pod \"community-operators-ppdjl\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.376171 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzwg4\" (UniqueName: \"kubernetes.io/projected/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-kube-api-access-bzwg4\") pod \"community-operators-ppdjl\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.376658 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-utilities\") pod \"community-operators-ppdjl\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.479186 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-catalog-content\") pod \"community-operators-ppdjl\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.479276 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzwg4\" (UniqueName: \"kubernetes.io/projected/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-kube-api-access-bzwg4\") pod \"community-operators-ppdjl\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.479352 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-utilities\") pod \"community-operators-ppdjl\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.479925 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-utilities\") pod \"community-operators-ppdjl\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.480045 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-catalog-content\") pod \"community-operators-ppdjl\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.503980 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzwg4\" (UniqueName: \"kubernetes.io/projected/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-kube-api-access-bzwg4\") pod \"community-operators-ppdjl\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:34 crc kubenswrapper[4936]: I0320 16:26:34.629460 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:35 crc kubenswrapper[4936]: I0320 16:26:35.225091 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppdjl"] Mar 20 16:26:35 crc kubenswrapper[4936]: I0320 16:26:35.541018 4936 generic.go:334] "Generic (PLEG): container finished" podID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerID="418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9" exitCode=0 Mar 20 16:26:35 crc kubenswrapper[4936]: I0320 16:26:35.541068 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppdjl" event={"ID":"b785e135-b3f2-4e1c-a5c2-bab595ea4fde","Type":"ContainerDied","Data":"418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9"} Mar 20 16:26:35 crc kubenswrapper[4936]: I0320 16:26:35.541096 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppdjl" event={"ID":"b785e135-b3f2-4e1c-a5c2-bab595ea4fde","Type":"ContainerStarted","Data":"1d6e6a24b27a471988c19635aa496d0baeebb93c01b102cd64a7cc2185b7e78e"} Mar 20 16:26:37 crc kubenswrapper[4936]: I0320 16:26:37.559639 4936 generic.go:334] "Generic (PLEG): container finished" podID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerID="589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac" exitCode=0 Mar 20 16:26:37 crc kubenswrapper[4936]: I0320 16:26:37.559737 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppdjl" event={"ID":"b785e135-b3f2-4e1c-a5c2-bab595ea4fde","Type":"ContainerDied","Data":"589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac"} Mar 20 16:26:38 crc kubenswrapper[4936]: I0320 16:26:38.575147 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppdjl" event={"ID":"b785e135-b3f2-4e1c-a5c2-bab595ea4fde","Type":"ContainerStarted","Data":"4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67"} Mar 20 16:26:38 crc kubenswrapper[4936]: I0320 16:26:38.594018 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ppdjl" podStartSLOduration=2.131392439 podStartE2EDuration="4.593999701s" podCreationTimestamp="2026-03-20 16:26:34 +0000 UTC" firstStartedPulling="2026-03-20 16:26:35.543298223 +0000 UTC m=+1546.489666038" lastFinishedPulling="2026-03-20 16:26:38.005905485 +0000 UTC m=+1548.952273300" observedRunningTime="2026-03-20 16:26:38.593716273 +0000 UTC m=+1549.540084108" watchObservedRunningTime="2026-03-20 16:26:38.593999701 +0000 UTC m=+1549.540367516" Mar 20 16:26:41 crc kubenswrapper[4936]: I0320 16:26:41.240418 4936 scope.go:117] "RemoveContainer" containerID="bc0e47a1192a0d56c16fbbc84a264cfcc6b0bcb470db43b06e0788b81d26b633" Mar 20 16:26:44 crc kubenswrapper[4936]: I0320 16:26:44.629809 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:44 crc kubenswrapper[4936]: I0320 16:26:44.630404 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:44 crc kubenswrapper[4936]: I0320 16:26:44.685750 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:45 crc kubenswrapper[4936]: I0320 16:26:45.695434 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:45 crc kubenswrapper[4936]: I0320 16:26:45.749682 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppdjl"] Mar 20 16:26:47 crc kubenswrapper[4936]: I0320 16:26:47.665318 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ppdjl" podUID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerName="registry-server" containerID="cri-o://4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67" gracePeriod=2 Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.403658 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.481126 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-utilities\") pod \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.481290 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzwg4\" (UniqueName: \"kubernetes.io/projected/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-kube-api-access-bzwg4\") pod \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.481326 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-catalog-content\") pod \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\" (UID: \"b785e135-b3f2-4e1c-a5c2-bab595ea4fde\") " Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.482054 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-utilities" (OuterVolumeSpecName: "utilities") pod "b785e135-b3f2-4e1c-a5c2-bab595ea4fde" (UID: "b785e135-b3f2-4e1c-a5c2-bab595ea4fde"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.482582 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.486513 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-kube-api-access-bzwg4" (OuterVolumeSpecName: "kube-api-access-bzwg4") pod "b785e135-b3f2-4e1c-a5c2-bab595ea4fde" (UID: "b785e135-b3f2-4e1c-a5c2-bab595ea4fde"). InnerVolumeSpecName "kube-api-access-bzwg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.546443 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b785e135-b3f2-4e1c-a5c2-bab595ea4fde" (UID: "b785e135-b3f2-4e1c-a5c2-bab595ea4fde"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.584797 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzwg4\" (UniqueName: \"kubernetes.io/projected/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-kube-api-access-bzwg4\") on node \"crc\" DevicePath \"\"" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.584832 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b785e135-b3f2-4e1c-a5c2-bab595ea4fde-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.675262 4936 generic.go:334] "Generic (PLEG): container finished" podID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerID="4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67" exitCode=0 Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.675331 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppdjl" event={"ID":"b785e135-b3f2-4e1c-a5c2-bab595ea4fde","Type":"ContainerDied","Data":"4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67"} Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.675363 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppdjl" event={"ID":"b785e135-b3f2-4e1c-a5c2-bab595ea4fde","Type":"ContainerDied","Data":"1d6e6a24b27a471988c19635aa496d0baeebb93c01b102cd64a7cc2185b7e78e"} Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.675383 4936 scope.go:117] "RemoveContainer" containerID="4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.675659 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppdjl" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.705954 4936 scope.go:117] "RemoveContainer" containerID="589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.718141 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppdjl"] Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.726824 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ppdjl"] Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.737810 4936 scope.go:117] "RemoveContainer" containerID="418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.769486 4936 scope.go:117] "RemoveContainer" containerID="4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67" Mar 20 16:26:48 crc kubenswrapper[4936]: E0320 16:26:48.770123 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67\": container with ID starting with 4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67 not found: ID does not exist" containerID="4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.770178 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67"} err="failed to get container status \"4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67\": rpc error: code = NotFound desc = could not find container \"4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67\": container with ID starting with 4a4c37480283960f6f2a4e770eb33bbfe51905d359cd00140fd5dcbb9c9e7a67 not found: ID does not exist" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.770213 4936 scope.go:117] "RemoveContainer" containerID="589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac" Mar 20 16:26:48 crc kubenswrapper[4936]: E0320 16:26:48.770621 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac\": container with ID starting with 589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac not found: ID does not exist" containerID="589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.770664 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac"} err="failed to get container status \"589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac\": rpc error: code = NotFound desc = could not find container \"589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac\": container with ID starting with 589cb0bc499dcbd468d9a2fb4a3925c0dd7cd4075223f5bbac45c662daf320ac not found: ID does not exist" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.770680 4936 scope.go:117] "RemoveContainer" containerID="418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9" Mar 20 16:26:48 crc kubenswrapper[4936]: E0320 16:26:48.770953 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9\": container with ID starting with 418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9 not found: ID does not exist" containerID="418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9" Mar 20 16:26:48 crc kubenswrapper[4936]: I0320 16:26:48.770995 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9"} err="failed to get container status \"418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9\": rpc error: code = NotFound desc = could not find container \"418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9\": container with ID starting with 418f874d6dbe080cace6acf61b674df4f9cff4369205d612667071d93beca1c9 not found: ID does not exist" Mar 20 16:26:49 crc kubenswrapper[4936]: I0320 16:26:49.600746 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-f9gjz_0df86e9a-2b40-4705-9c88-2f24b1037d9d/kube-rbac-proxy/0.log" Mar 20 16:26:49 crc kubenswrapper[4936]: I0320 16:26:49.721978 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-f9gjz_0df86e9a-2b40-4705-9c88-2f24b1037d9d/controller/0.log" Mar 20 16:26:49 crc kubenswrapper[4936]: I0320 16:26:49.851293 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-frr-files/0.log" Mar 20 16:26:49 crc kubenswrapper[4936]: I0320 16:26:49.877658 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" path="/var/lib/kubelet/pods/b785e135-b3f2-4e1c-a5c2-bab595ea4fde/volumes" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.052036 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-reloader/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.121530 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-metrics/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.132314 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-frr-files/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.146174 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-reloader/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.320774 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-frr-files/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.381184 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-reloader/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.433798 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-metrics/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.433805 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-metrics/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.564592 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-frr-files/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.587534 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-reloader/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.635909 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/cp-metrics/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.656406 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/controller/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.789840 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/frr-metrics/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.858049 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/kube-rbac-proxy/0.log" Mar 20 16:26:50 crc kubenswrapper[4936]: I0320 16:26:50.872041 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/kube-rbac-proxy-frr/0.log" Mar 20 16:26:51 crc kubenswrapper[4936]: I0320 16:26:51.008197 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/reloader/0.log" Mar 20 16:26:51 crc kubenswrapper[4936]: I0320 16:26:51.090934 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-5tb4t_94012e02-d626-4c3c-b57c-c62e05f7e5dd/frr-k8s-webhook-server/0.log" Mar 20 16:26:51 crc kubenswrapper[4936]: I0320 16:26:51.248350 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7db57c7955-sxrb7_b880eadb-7994-485a-99fc-f7a4265707d3/manager/0.log" Mar 20 16:26:51 crc kubenswrapper[4936]: I0320 16:26:51.472090 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b8d4cf8dd-js5nz_05880ff9-4310-40a0-a8f3-47dfc79f87e2/webhook-server/0.log" Mar 20 16:26:51 crc kubenswrapper[4936]: I0320 16:26:51.621077 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r7c5t_bac69582-9f6c-4828-a05b-1f9bfbcefbef/kube-rbac-proxy/0.log" Mar 20 16:26:51 crc kubenswrapper[4936]: I0320 16:26:51.796227 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4fg6x_b2b10075-38e1-46a0-a69e-f076348c77c5/frr/0.log" Mar 20 16:26:52 crc kubenswrapper[4936]: I0320 16:26:52.023001 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r7c5t_bac69582-9f6c-4828-a05b-1f9bfbcefbef/speaker/0.log" Mar 20 16:26:58 crc kubenswrapper[4936]: I0320 16:26:58.854736 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:26:58 crc kubenswrapper[4936]: I0320 16:26:58.855301 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:26:58 crc kubenswrapper[4936]: I0320 16:26:58.855357 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:26:58 crc kubenswrapper[4936]: I0320 16:26:58.856061 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f54448fa33df6b9490f8d0906969969ecfdaf5a9c74fdcedd6ab26b4c513d8e"} pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 16:26:58 crc kubenswrapper[4936]: I0320 16:26:58.856153 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" containerID="cri-o://7f54448fa33df6b9490f8d0906969969ecfdaf5a9c74fdcedd6ab26b4c513d8e" gracePeriod=600 Mar 20 16:26:59 crc kubenswrapper[4936]: I0320 16:26:59.791426 4936 generic.go:334] "Generic (PLEG): container finished" podID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerID="7f54448fa33df6b9490f8d0906969969ecfdaf5a9c74fdcedd6ab26b4c513d8e" exitCode=0 Mar 20 16:26:59 crc kubenswrapper[4936]: I0320 16:26:59.791492 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerDied","Data":"7f54448fa33df6b9490f8d0906969969ecfdaf5a9c74fdcedd6ab26b4c513d8e"} Mar 20 16:26:59 crc kubenswrapper[4936]: I0320 16:26:59.792537 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerStarted","Data":"718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a"} Mar 20 16:26:59 crc kubenswrapper[4936]: I0320 16:26:59.792603 4936 scope.go:117] "RemoveContainer" containerID="19b7ae40534ebc0df1ca16d7a38e73567bfe3ddd0d163640d422629548acfbe8" Mar 20 16:27:04 crc kubenswrapper[4936]: I0320 16:27:04.775752 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm_88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a/util/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.050312 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm_88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a/util/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.066767 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm_88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a/pull/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.066985 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm_88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a/pull/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.210804 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm_88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a/util/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.226348 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm_88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a/pull/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.271980 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rt5rm_88ecb5d1-7f00-450d-a93f-7ea8b50d0d0a/extract/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.404032 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj_ece6daad-afcb-4bfb-ace9-1fb25a6c9f10/util/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.584060 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj_ece6daad-afcb-4bfb-ace9-1fb25a6c9f10/pull/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.584259 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj_ece6daad-afcb-4bfb-ace9-1fb25a6c9f10/pull/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.608300 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj_ece6daad-afcb-4bfb-ace9-1fb25a6c9f10/util/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.761236 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj_ece6daad-afcb-4bfb-ace9-1fb25a6c9f10/util/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.762369 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj_ece6daad-afcb-4bfb-ace9-1fb25a6c9f10/extract/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.775573 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f9cdj_ece6daad-afcb-4bfb-ace9-1fb25a6c9f10/pull/0.log" Mar 20 16:27:05 crc kubenswrapper[4936]: I0320 16:27:05.910861 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h6c8f_1399ef65-8fa1-48a3-b7a2-c7ebd25198e3/extract-utilities/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.092349 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h6c8f_1399ef65-8fa1-48a3-b7a2-c7ebd25198e3/extract-content/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.107300 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h6c8f_1399ef65-8fa1-48a3-b7a2-c7ebd25198e3/extract-utilities/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.136634 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h6c8f_1399ef65-8fa1-48a3-b7a2-c7ebd25198e3/extract-content/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.296862 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h6c8f_1399ef65-8fa1-48a3-b7a2-c7ebd25198e3/extract-utilities/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.300999 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h6c8f_1399ef65-8fa1-48a3-b7a2-c7ebd25198e3/extract-content/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.490090 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h6c8f_1399ef65-8fa1-48a3-b7a2-c7ebd25198e3/registry-server/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.522659 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-84rkf_043c1af0-d920-45f8-b48c-fdc26c752e9a/extract-utilities/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.699890 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-84rkf_043c1af0-d920-45f8-b48c-fdc26c752e9a/extract-content/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.728121 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-84rkf_043c1af0-d920-45f8-b48c-fdc26c752e9a/extract-content/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.732015 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-84rkf_043c1af0-d920-45f8-b48c-fdc26c752e9a/extract-utilities/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.917779 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-84rkf_043c1af0-d920-45f8-b48c-fdc26c752e9a/extract-utilities/0.log" Mar 20 16:27:06 crc kubenswrapper[4936]: I0320 16:27:06.933442 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-84rkf_043c1af0-d920-45f8-b48c-fdc26c752e9a/extract-content/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.214106 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ldm5b_d82b61cf-0c16-48b7-9a4f-c42b9212d85a/marketplace-operator/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.303875 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tncvv_bb0083ee-8688-481a-86c6-df9338ea1364/extract-utilities/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.311963 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-84rkf_043c1af0-d920-45f8-b48c-fdc26c752e9a/registry-server/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.515532 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tncvv_bb0083ee-8688-481a-86c6-df9338ea1364/extract-utilities/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.515637 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tncvv_bb0083ee-8688-481a-86c6-df9338ea1364/extract-content/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.526061 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tncvv_bb0083ee-8688-481a-86c6-df9338ea1364/extract-content/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.705116 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tncvv_bb0083ee-8688-481a-86c6-df9338ea1364/extract-utilities/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.726135 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tncvv_bb0083ee-8688-481a-86c6-df9338ea1364/extract-content/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.798372 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tncvv_bb0083ee-8688-481a-86c6-df9338ea1364/registry-server/0.log" Mar 20 16:27:07 crc kubenswrapper[4936]: I0320 16:27:07.906476 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lq9hg_51fc6176-8334-4c16-bfd3-42584385ea47/extract-utilities/0.log" Mar 20 16:27:08 crc kubenswrapper[4936]: I0320 16:27:08.025350 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lq9hg_51fc6176-8334-4c16-bfd3-42584385ea47/extract-content/0.log" Mar 20 16:27:08 crc kubenswrapper[4936]: I0320 16:27:08.026830 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lq9hg_51fc6176-8334-4c16-bfd3-42584385ea47/extract-utilities/0.log" Mar 20 16:27:08 crc kubenswrapper[4936]: I0320 16:27:08.058669 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lq9hg_51fc6176-8334-4c16-bfd3-42584385ea47/extract-content/0.log" Mar 20 16:27:08 crc kubenswrapper[4936]: I0320 16:27:08.272414 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lq9hg_51fc6176-8334-4c16-bfd3-42584385ea47/extract-content/0.log" Mar 20 16:27:08 crc kubenswrapper[4936]: I0320 16:27:08.292620 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lq9hg_51fc6176-8334-4c16-bfd3-42584385ea47/extract-utilities/0.log" Mar 20 16:27:08 crc kubenswrapper[4936]: I0320 16:27:08.562225 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lq9hg_51fc6176-8334-4c16-bfd3-42584385ea47/registry-server/0.log" Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.054139 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7948-account-create-update-8trc2"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.064112 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-05f5-account-create-update-9vd8r"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.075932 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-jvq45"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.087697 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-zphvf"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.096002 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-962c-account-create-update-nbrwz"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.103314 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7948-account-create-update-8trc2"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.110572 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-zphvf"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.117979 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-962c-account-create-update-nbrwz"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.125505 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-jvq45"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.133216 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-9vpbv"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.140324 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-05f5-account-create-update-9vd8r"] Mar 20 16:27:58 crc kubenswrapper[4936]: I0320 16:27:58.147357 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-9vpbv"] Mar 20 16:27:59 crc kubenswrapper[4936]: I0320 16:27:59.865938 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11a87bf1-c0b8-47a2-8ccb-faf322bebfb1" path="/var/lib/kubelet/pods/11a87bf1-c0b8-47a2-8ccb-faf322bebfb1/volumes" Mar 20 16:27:59 crc kubenswrapper[4936]: I0320 16:27:59.866984 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434fff7b-3810-45b8-8028-98a9d77d3f15" path="/var/lib/kubelet/pods/434fff7b-3810-45b8-8028-98a9d77d3f15/volumes" Mar 20 16:27:59 crc kubenswrapper[4936]: I0320 16:27:59.867628 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="812b9782-84b1-41fc-939e-30e82c1b5e13" path="/var/lib/kubelet/pods/812b9782-84b1-41fc-939e-30e82c1b5e13/volumes" Mar 20 16:27:59 crc kubenswrapper[4936]: I0320 16:27:59.868379 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0aa18c2-d52b-4671-996b-985c848a748f" path="/var/lib/kubelet/pods/a0aa18c2-d52b-4671-996b-985c848a748f/volumes" Mar 20 16:27:59 crc kubenswrapper[4936]: I0320 16:27:59.869936 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8df2114-9d0b-43c8-91cd-a60355f7238b" path="/var/lib/kubelet/pods/b8df2114-9d0b-43c8-91cd-a60355f7238b/volumes" Mar 20 16:27:59 crc kubenswrapper[4936]: I0320 16:27:59.870913 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48f3728-76d9-43c5-8053-b27ac3751ecb" path="/var/lib/kubelet/pods/c48f3728-76d9-43c5-8053-b27ac3751ecb/volumes" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.143683 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567068-vfxgs"] Mar 20 16:28:00 crc kubenswrapper[4936]: E0320 16:28:00.144099 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerName="registry-server" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.144134 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerName="registry-server" Mar 20 16:28:00 crc kubenswrapper[4936]: E0320 16:28:00.144153 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerName="extract-content" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.144160 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerName="extract-content" Mar 20 16:28:00 crc kubenswrapper[4936]: E0320 16:28:00.144174 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerName="extract-utilities" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.144180 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerName="extract-utilities" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.144428 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="b785e135-b3f2-4e1c-a5c2-bab595ea4fde" containerName="registry-server" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.145125 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567068-vfxgs" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.147084 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.149269 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.149419 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.153420 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567068-vfxgs"] Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.274314 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq7md\" (UniqueName: \"kubernetes.io/projected/2c5319ea-6340-41c9-ae05-0ecfed359126-kube-api-access-jq7md\") pod \"auto-csr-approver-29567068-vfxgs\" (UID: \"2c5319ea-6340-41c9-ae05-0ecfed359126\") " pod="openshift-infra/auto-csr-approver-29567068-vfxgs" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.375568 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq7md\" (UniqueName: \"kubernetes.io/projected/2c5319ea-6340-41c9-ae05-0ecfed359126-kube-api-access-jq7md\") pod \"auto-csr-approver-29567068-vfxgs\" (UID: \"2c5319ea-6340-41c9-ae05-0ecfed359126\") " pod="openshift-infra/auto-csr-approver-29567068-vfxgs" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.393261 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq7md\" (UniqueName: \"kubernetes.io/projected/2c5319ea-6340-41c9-ae05-0ecfed359126-kube-api-access-jq7md\") pod \"auto-csr-approver-29567068-vfxgs\" (UID: \"2c5319ea-6340-41c9-ae05-0ecfed359126\") " pod="openshift-infra/auto-csr-approver-29567068-vfxgs" Mar 20 16:28:00 crc kubenswrapper[4936]: I0320 16:28:00.472452 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567068-vfxgs" Mar 20 16:28:01 crc kubenswrapper[4936]: I0320 16:28:01.061595 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567068-vfxgs"] Mar 20 16:28:01 crc kubenswrapper[4936]: I0320 16:28:01.319425 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567068-vfxgs" event={"ID":"2c5319ea-6340-41c9-ae05-0ecfed359126","Type":"ContainerStarted","Data":"1cfe5e8a45941dd82b36cd8b222eb801e1c002d003d54c8262ad20065e2d49fd"} Mar 20 16:28:02 crc kubenswrapper[4936]: I0320 16:28:02.034931 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-cjk9m"] Mar 20 16:28:02 crc kubenswrapper[4936]: I0320 16:28:02.047958 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-cjk9m"] Mar 20 16:28:03 crc kubenswrapper[4936]: I0320 16:28:03.350864 4936 generic.go:334] "Generic (PLEG): container finished" podID="2c5319ea-6340-41c9-ae05-0ecfed359126" containerID="32dfe8698664260bd88bc6ae5d335de970fcf1c4f8bdb1fed2084dea8979a23c" exitCode=0 Mar 20 16:28:03 crc kubenswrapper[4936]: I0320 16:28:03.351181 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567068-vfxgs" event={"ID":"2c5319ea-6340-41c9-ae05-0ecfed359126","Type":"ContainerDied","Data":"32dfe8698664260bd88bc6ae5d335de970fcf1c4f8bdb1fed2084dea8979a23c"} Mar 20 16:28:03 crc kubenswrapper[4936]: I0320 16:28:03.869451 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aece10e-0797-48c2-aac7-7a4b12cf0cf5" path="/var/lib/kubelet/pods/3aece10e-0797-48c2-aac7-7a4b12cf0cf5/volumes" Mar 20 16:28:04 crc kubenswrapper[4936]: I0320 16:28:04.758122 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567068-vfxgs" Mar 20 16:28:04 crc kubenswrapper[4936]: I0320 16:28:04.784538 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq7md\" (UniqueName: \"kubernetes.io/projected/2c5319ea-6340-41c9-ae05-0ecfed359126-kube-api-access-jq7md\") pod \"2c5319ea-6340-41c9-ae05-0ecfed359126\" (UID: \"2c5319ea-6340-41c9-ae05-0ecfed359126\") " Mar 20 16:28:04 crc kubenswrapper[4936]: I0320 16:28:04.790225 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c5319ea-6340-41c9-ae05-0ecfed359126-kube-api-access-jq7md" (OuterVolumeSpecName: "kube-api-access-jq7md") pod "2c5319ea-6340-41c9-ae05-0ecfed359126" (UID: "2c5319ea-6340-41c9-ae05-0ecfed359126"). InnerVolumeSpecName "kube-api-access-jq7md". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:28:04 crc kubenswrapper[4936]: I0320 16:28:04.886300 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq7md\" (UniqueName: \"kubernetes.io/projected/2c5319ea-6340-41c9-ae05-0ecfed359126-kube-api-access-jq7md\") on node \"crc\" DevicePath \"\"" Mar 20 16:28:05 crc kubenswrapper[4936]: I0320 16:28:05.377606 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567068-vfxgs" event={"ID":"2c5319ea-6340-41c9-ae05-0ecfed359126","Type":"ContainerDied","Data":"1cfe5e8a45941dd82b36cd8b222eb801e1c002d003d54c8262ad20065e2d49fd"} Mar 20 16:28:05 crc kubenswrapper[4936]: I0320 16:28:05.377634 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567068-vfxgs" Mar 20 16:28:05 crc kubenswrapper[4936]: I0320 16:28:05.377645 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cfe5e8a45941dd82b36cd8b222eb801e1c002d003d54c8262ad20065e2d49fd" Mar 20 16:28:05 crc kubenswrapper[4936]: I0320 16:28:05.823819 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567062-b6frx"] Mar 20 16:28:05 crc kubenswrapper[4936]: I0320 16:28:05.834169 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567062-b6frx"] Mar 20 16:28:05 crc kubenswrapper[4936]: I0320 16:28:05.867438 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="643cf02c-71a1-498a-924d-ff539faf7bc7" path="/var/lib/kubelet/pods/643cf02c-71a1-498a-924d-ff539faf7bc7/volumes" Mar 20 16:28:31 crc kubenswrapper[4936]: I0320 16:28:31.045836 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-sg9df"] Mar 20 16:28:31 crc kubenswrapper[4936]: I0320 16:28:31.054100 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-sg9df"] Mar 20 16:28:31 crc kubenswrapper[4936]: I0320 16:28:31.866438 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a817ffe2-32a4-4c37-9342-5d642d179ea7" path="/var/lib/kubelet/pods/a817ffe2-32a4-4c37-9342-5d642d179ea7/volumes" Mar 20 16:28:41 crc kubenswrapper[4936]: I0320 16:28:41.390843 4936 scope.go:117] "RemoveContainer" containerID="3d2fce81bfbcdbdd93bcb9cfb855367cdb97743b31bb2d69a5ea82518876816f" Mar 20 16:28:41 crc kubenswrapper[4936]: I0320 16:28:41.452022 4936 scope.go:117] "RemoveContainer" containerID="74158cac144ee91c9dfde123deaa764b39ca9122d5eaaf455154f644623db967" Mar 20 16:28:41 crc kubenswrapper[4936]: I0320 16:28:41.470872 4936 scope.go:117] "RemoveContainer" containerID="2f49546a1d1fa1324c648ba58556f65bb5b210cdb9ce4ff81c0f4e813897788d" Mar 20 16:28:41 crc kubenswrapper[4936]: I0320 16:28:41.520640 4936 scope.go:117] "RemoveContainer" containerID="9d1d5fa5c224fda226d2cb9276b7341cf1d644f1375cb62c6e65bddb1a799120" Mar 20 16:28:41 crc kubenswrapper[4936]: I0320 16:28:41.570534 4936 scope.go:117] "RemoveContainer" containerID="4e22f1f602d60c236a7c02e8361dd0b77f3c35628b6dc95b2ae098e884f08b6e" Mar 20 16:28:41 crc kubenswrapper[4936]: I0320 16:28:41.628024 4936 scope.go:117] "RemoveContainer" containerID="2d73a47e2620619edcff2f7a567e03b95fd0efb034c602109094656476739fc1" Mar 20 16:28:41 crc kubenswrapper[4936]: I0320 16:28:41.672982 4936 scope.go:117] "RemoveContainer" containerID="666a52a4fc6ae392ee80e61155687b982c432fe0f44003898f24d68084097b76" Mar 20 16:28:41 crc kubenswrapper[4936]: I0320 16:28:41.695673 4936 scope.go:117] "RemoveContainer" containerID="f52edbe07ed941a85bc6a10d310a58fcaeb3c1ad929a5efa5f9ee8810ad218f9" Mar 20 16:28:41 crc kubenswrapper[4936]: I0320 16:28:41.717289 4936 scope.go:117] "RemoveContainer" containerID="75f5f9b53ac503449b410ff96fd3effa3da09f9333ba2c1c601660b6f5a88afb" Mar 20 16:28:45 crc kubenswrapper[4936]: I0320 16:28:45.797912 4936 generic.go:334] "Generic (PLEG): container finished" podID="631b69e4-abc4-431f-9edb-d076726545e7" containerID="937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e" exitCode=0 Mar 20 16:28:45 crc kubenswrapper[4936]: I0320 16:28:45.798005 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xpdzx/must-gather-glbch" event={"ID":"631b69e4-abc4-431f-9edb-d076726545e7","Type":"ContainerDied","Data":"937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e"} Mar 20 16:28:45 crc kubenswrapper[4936]: I0320 16:28:45.798757 4936 scope.go:117] "RemoveContainer" containerID="937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e" Mar 20 16:28:46 crc kubenswrapper[4936]: I0320 16:28:46.798221 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xpdzx_must-gather-glbch_631b69e4-abc4-431f-9edb-d076726545e7/gather/0.log" Mar 20 16:28:51 crc kubenswrapper[4936]: I0320 16:28:51.026934 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-p249h"] Mar 20 16:28:51 crc kubenswrapper[4936]: I0320 16:28:51.040072 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-p249h"] Mar 20 16:28:51 crc kubenswrapper[4936]: I0320 16:28:51.868493 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2726b8ff-699c-419e-ae52-0868c355ea23" path="/var/lib/kubelet/pods/2726b8ff-699c-419e-ae52-0868c355ea23/volumes" Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.249378 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xpdzx/must-gather-glbch"] Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.250029 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xpdzx/must-gather-glbch" podUID="631b69e4-abc4-431f-9edb-d076726545e7" containerName="copy" containerID="cri-o://12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9" gracePeriod=2 Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.259683 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xpdzx/must-gather-glbch"] Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.679005 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xpdzx_must-gather-glbch_631b69e4-abc4-431f-9edb-d076726545e7/copy/0.log" Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.679778 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.869267 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/631b69e4-abc4-431f-9edb-d076726545e7-must-gather-output\") pod \"631b69e4-abc4-431f-9edb-d076726545e7\" (UID: \"631b69e4-abc4-431f-9edb-d076726545e7\") " Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.869651 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pzwj\" (UniqueName: \"kubernetes.io/projected/631b69e4-abc4-431f-9edb-d076726545e7-kube-api-access-7pzwj\") pod \"631b69e4-abc4-431f-9edb-d076726545e7\" (UID: \"631b69e4-abc4-431f-9edb-d076726545e7\") " Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.876051 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631b69e4-abc4-431f-9edb-d076726545e7-kube-api-access-7pzwj" (OuterVolumeSpecName: "kube-api-access-7pzwj") pod "631b69e4-abc4-431f-9edb-d076726545e7" (UID: "631b69e4-abc4-431f-9edb-d076726545e7"). InnerVolumeSpecName "kube-api-access-7pzwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.887482 4936 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xpdzx_must-gather-glbch_631b69e4-abc4-431f-9edb-d076726545e7/copy/0.log" Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.887861 4936 generic.go:334] "Generic (PLEG): container finished" podID="631b69e4-abc4-431f-9edb-d076726545e7" containerID="12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9" exitCode=143 Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.887908 4936 scope.go:117] "RemoveContainer" containerID="12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9" Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.888042 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xpdzx/must-gather-glbch" Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.943519 4936 scope.go:117] "RemoveContainer" containerID="937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e" Mar 20 16:28:54 crc kubenswrapper[4936]: I0320 16:28:54.975519 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pzwj\" (UniqueName: \"kubernetes.io/projected/631b69e4-abc4-431f-9edb-d076726545e7-kube-api-access-7pzwj\") on node \"crc\" DevicePath \"\"" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.043769 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-s46fs"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.048882 4936 scope.go:117] "RemoveContainer" containerID="12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9" Mar 20 16:28:55 crc kubenswrapper[4936]: E0320 16:28:55.049486 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9\": container with ID starting with 12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9 not found: ID does not exist" containerID="12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.049531 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9"} err="failed to get container status \"12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9\": rpc error: code = NotFound desc = could not find container \"12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9\": container with ID starting with 12cb526c13d7ce7690518cd5b720599617ea3ce10ff8283994f59a392d5976d9 not found: ID does not exist" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.049589 4936 scope.go:117] "RemoveContainer" containerID="937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e" Mar 20 16:28:55 crc kubenswrapper[4936]: E0320 16:28:55.050737 4936 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e\": container with ID starting with 937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e not found: ID does not exist" containerID="937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.050784 4936 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e"} err="failed to get container status \"937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e\": rpc error: code = NotFound desc = could not find container \"937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e\": container with ID starting with 937d0c874497c31a2718fcf0eed63c0a3d167e6036ccb2e9e22d6d6ad373e50e not found: ID does not exist" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.051636 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/631b69e4-abc4-431f-9edb-d076726545e7-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "631b69e4-abc4-431f-9edb-d076726545e7" (UID: "631b69e4-abc4-431f-9edb-d076726545e7"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.059683 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-s46fs"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.068396 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-m5pnp"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.076282 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7437-account-create-update-qn4hp"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.077071 4936 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/631b69e4-abc4-431f-9edb-d076726545e7-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.084746 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-63a4-account-create-update-wng8g"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.093704 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7437-account-create-update-qn4hp"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.102035 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-m5pnp"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.109592 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ab53-account-create-update-rrk7z"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.117215 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-63a4-account-create-update-wng8g"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.124608 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ab53-account-create-update-rrk7z"] Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.867789 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0288b360-6ff9-4568-bc31-1351a3e55ad2" path="/var/lib/kubelet/pods/0288b360-6ff9-4568-bc31-1351a3e55ad2/volumes" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.868915 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="631b69e4-abc4-431f-9edb-d076726545e7" path="/var/lib/kubelet/pods/631b69e4-abc4-431f-9edb-d076726545e7/volumes" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.870122 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b802c0e2-15be-4604-b64c-3b7441a1096d" path="/var/lib/kubelet/pods/b802c0e2-15be-4604-b64c-3b7441a1096d/volumes" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.872974 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da80eafb-7c87-4b87-89ae-beee42d9b256" path="/var/lib/kubelet/pods/da80eafb-7c87-4b87-89ae-beee42d9b256/volumes" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.874028 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df4dcd90-70c5-4d83-b8a9-db6d914a8a2d" path="/var/lib/kubelet/pods/df4dcd90-70c5-4d83-b8a9-db6d914a8a2d/volumes" Mar 20 16:28:55 crc kubenswrapper[4936]: I0320 16:28:55.874710 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df8f8d6e-b8f0-4148-940e-712399f97ab1" path="/var/lib/kubelet/pods/df8f8d6e-b8f0-4148-940e-712399f97ab1/volumes" Mar 20 16:28:59 crc kubenswrapper[4936]: I0320 16:28:59.029453 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wfxxc"] Mar 20 16:28:59 crc kubenswrapper[4936]: I0320 16:28:59.039119 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wfxxc"] Mar 20 16:28:59 crc kubenswrapper[4936]: I0320 16:28:59.871832 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e13cc3e4-9442-40ec-b43f-cb968e26a0e1" path="/var/lib/kubelet/pods/e13cc3e4-9442-40ec-b43f-cb968e26a0e1/volumes" Mar 20 16:29:28 crc kubenswrapper[4936]: I0320 16:29:28.854750 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:29:28 crc kubenswrapper[4936]: I0320 16:29:28.855418 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:29:41 crc kubenswrapper[4936]: I0320 16:29:41.915663 4936 scope.go:117] "RemoveContainer" containerID="6708d423b696ff66673a16646000716783814ae4031db65a5f25877b13a8bae9" Mar 20 16:29:41 crc kubenswrapper[4936]: I0320 16:29:41.966596 4936 scope.go:117] "RemoveContainer" containerID="29a8dda96b57bda5eebb47e6d6bc02216ba11a1b882a776df41be2834c8291e0" Mar 20 16:29:42 crc kubenswrapper[4936]: I0320 16:29:42.029775 4936 scope.go:117] "RemoveContainer" containerID="3aa99a195b752614ecace8abbec25d0c2e4d058279b0d984fecc8d8198321f9c" Mar 20 16:29:42 crc kubenswrapper[4936]: I0320 16:29:42.072146 4936 scope.go:117] "RemoveContainer" containerID="e2964864b07f763a8cdc75e13ac493e4e214e805a94f28ce5767295b5474e76b" Mar 20 16:29:42 crc kubenswrapper[4936]: I0320 16:29:42.118706 4936 scope.go:117] "RemoveContainer" containerID="89fdffa55aba7f23c2c414ea31524a469dd7b82ebff0488389910f625da859c4" Mar 20 16:29:42 crc kubenswrapper[4936]: I0320 16:29:42.142309 4936 scope.go:117] "RemoveContainer" containerID="0b6794161057d16f86c7ab3b7cfa51d5923f92378407959f1d03bfd9c2857f77" Mar 20 16:29:42 crc kubenswrapper[4936]: I0320 16:29:42.187892 4936 scope.go:117] "RemoveContainer" containerID="54a706b4981c4361f3f965a7f06842c315d8cf05db79fa68b6a6fb3dc232c85a" Mar 20 16:29:50 crc kubenswrapper[4936]: I0320 16:29:50.052116 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-w6hwd"] Mar 20 16:29:50 crc kubenswrapper[4936]: I0320 16:29:50.066664 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4t65w"] Mar 20 16:29:50 crc kubenswrapper[4936]: I0320 16:29:50.077007 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-w6hwd"] Mar 20 16:29:50 crc kubenswrapper[4936]: I0320 16:29:50.086983 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4t65w"] Mar 20 16:29:51 crc kubenswrapper[4936]: I0320 16:29:51.866086 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd7ee68-8aa9-4b2c-9d06-c4de9698b755" path="/var/lib/kubelet/pods/3cd7ee68-8aa9-4b2c-9d06-c4de9698b755/volumes" Mar 20 16:29:51 crc kubenswrapper[4936]: I0320 16:29:51.867223 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5421152c-d474-428f-9554-96e036719655" path="/var/lib/kubelet/pods/5421152c-d474-428f-9554-96e036719655/volumes" Mar 20 16:29:57 crc kubenswrapper[4936]: I0320 16:29:57.359714 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-74b5cc8579-wqhwv" podUID="9284dab5-bdd1-4792-8944-0f9c820a927d" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 20 16:29:58 crc kubenswrapper[4936]: I0320 16:29:58.854853 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:29:58 crc kubenswrapper[4936]: I0320 16:29:58.855179 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.179464 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5"] Mar 20 16:30:00 crc kubenswrapper[4936]: E0320 16:30:00.180797 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631b69e4-abc4-431f-9edb-d076726545e7" containerName="copy" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.180885 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="631b69e4-abc4-431f-9edb-d076726545e7" containerName="copy" Mar 20 16:30:00 crc kubenswrapper[4936]: E0320 16:30:00.180977 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631b69e4-abc4-431f-9edb-d076726545e7" containerName="gather" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.181030 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="631b69e4-abc4-431f-9edb-d076726545e7" containerName="gather" Mar 20 16:30:00 crc kubenswrapper[4936]: E0320 16:30:00.181085 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5319ea-6340-41c9-ae05-0ecfed359126" containerName="oc" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.181135 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5319ea-6340-41c9-ae05-0ecfed359126" containerName="oc" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.181401 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5319ea-6340-41c9-ae05-0ecfed359126" containerName="oc" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.181470 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="631b69e4-abc4-431f-9edb-d076726545e7" containerName="gather" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.181536 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="631b69e4-abc4-431f-9edb-d076726545e7" containerName="copy" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.182511 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.187042 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.187389 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.190646 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567070-ljgst"] Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.192271 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567070-ljgst" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.196832 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5"] Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.197721 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7cz4\" (UniqueName: \"kubernetes.io/projected/19b8c886-feae-44d0-a169-e07619ccad1b-kube-api-access-d7cz4\") pod \"auto-csr-approver-29567070-ljgst\" (UID: \"19b8c886-feae-44d0-a169-e07619ccad1b\") " pod="openshift-infra/auto-csr-approver-29567070-ljgst" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.197823 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7227e37e-faca-45f5-ba13-0877ea8fc688-secret-volume\") pod \"collect-profiles-29567070-252k5\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.198007 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7227e37e-faca-45f5-ba13-0877ea8fc688-config-volume\") pod \"collect-profiles-29567070-252k5\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.198194 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.198262 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvnxt\" (UniqueName: \"kubernetes.io/projected/7227e37e-faca-45f5-ba13-0877ea8fc688-kube-api-access-wvnxt\") pod \"collect-profiles-29567070-252k5\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.198217 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.200778 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.225953 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567070-ljgst"] Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.301386 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7227e37e-faca-45f5-ba13-0877ea8fc688-config-volume\") pod \"collect-profiles-29567070-252k5\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.301486 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvnxt\" (UniqueName: \"kubernetes.io/projected/7227e37e-faca-45f5-ba13-0877ea8fc688-kube-api-access-wvnxt\") pod \"collect-profiles-29567070-252k5\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.301706 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7cz4\" (UniqueName: \"kubernetes.io/projected/19b8c886-feae-44d0-a169-e07619ccad1b-kube-api-access-d7cz4\") pod \"auto-csr-approver-29567070-ljgst\" (UID: \"19b8c886-feae-44d0-a169-e07619ccad1b\") " pod="openshift-infra/auto-csr-approver-29567070-ljgst" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.302298 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7227e37e-faca-45f5-ba13-0877ea8fc688-secret-volume\") pod \"collect-profiles-29567070-252k5\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.302884 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7227e37e-faca-45f5-ba13-0877ea8fc688-config-volume\") pod \"collect-profiles-29567070-252k5\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.312697 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7227e37e-faca-45f5-ba13-0877ea8fc688-secret-volume\") pod \"collect-profiles-29567070-252k5\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.319996 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7cz4\" (UniqueName: \"kubernetes.io/projected/19b8c886-feae-44d0-a169-e07619ccad1b-kube-api-access-d7cz4\") pod \"auto-csr-approver-29567070-ljgst\" (UID: \"19b8c886-feae-44d0-a169-e07619ccad1b\") " pod="openshift-infra/auto-csr-approver-29567070-ljgst" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.323255 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvnxt\" (UniqueName: \"kubernetes.io/projected/7227e37e-faca-45f5-ba13-0877ea8fc688-kube-api-access-wvnxt\") pod \"collect-profiles-29567070-252k5\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.508881 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:00 crc kubenswrapper[4936]: I0320 16:30:00.525156 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567070-ljgst" Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.022865 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5"] Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.068338 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9w2gg"] Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.080417 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-j4d5s"] Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.092219 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-j4d5s"] Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.097962 4936 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.101636 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9w2gg"] Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.108327 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567070-ljgst"] Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.583955 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567070-ljgst" event={"ID":"19b8c886-feae-44d0-a169-e07619ccad1b","Type":"ContainerStarted","Data":"d3e5c3a3dc84eb9c2c8960d2bc6eda6a3d08dab98274b3ae3d1b2842a3cf5a6e"} Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.585881 4936 generic.go:334] "Generic (PLEG): container finished" podID="7227e37e-faca-45f5-ba13-0877ea8fc688" containerID="ecbb7300fc4e74d78ba904abf65b8e7b47027198cfcfa594a0181e21ea9c630b" exitCode=0 Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.585939 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" event={"ID":"7227e37e-faca-45f5-ba13-0877ea8fc688","Type":"ContainerDied","Data":"ecbb7300fc4e74d78ba904abf65b8e7b47027198cfcfa594a0181e21ea9c630b"} Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.585977 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" event={"ID":"7227e37e-faca-45f5-ba13-0877ea8fc688","Type":"ContainerStarted","Data":"60a1962f32d6c887d226ae0038120830d7ac705b4cce78122e0a56839829ea91"} Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.863892 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61aff63e-dc1b-4854-af4a-188837ad4cf9" path="/var/lib/kubelet/pods/61aff63e-dc1b-4854-af4a-188837ad4cf9/volumes" Mar 20 16:30:01 crc kubenswrapper[4936]: I0320 16:30:01.864525 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b87c4418-b13e-41e2-bbd8-d7dfb057a594" path="/var/lib/kubelet/pods/b87c4418-b13e-41e2-bbd8-d7dfb057a594/volumes" Mar 20 16:30:02 crc kubenswrapper[4936]: I0320 16:30:02.965273 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:02 crc kubenswrapper[4936]: I0320 16:30:02.967747 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7227e37e-faca-45f5-ba13-0877ea8fc688-config-volume\") pod \"7227e37e-faca-45f5-ba13-0877ea8fc688\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " Mar 20 16:30:02 crc kubenswrapper[4936]: I0320 16:30:02.967957 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7227e37e-faca-45f5-ba13-0877ea8fc688-secret-volume\") pod \"7227e37e-faca-45f5-ba13-0877ea8fc688\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " Mar 20 16:30:02 crc kubenswrapper[4936]: I0320 16:30:02.968036 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvnxt\" (UniqueName: \"kubernetes.io/projected/7227e37e-faca-45f5-ba13-0877ea8fc688-kube-api-access-wvnxt\") pod \"7227e37e-faca-45f5-ba13-0877ea8fc688\" (UID: \"7227e37e-faca-45f5-ba13-0877ea8fc688\") " Mar 20 16:30:02 crc kubenswrapper[4936]: I0320 16:30:02.970392 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7227e37e-faca-45f5-ba13-0877ea8fc688-config-volume" (OuterVolumeSpecName: "config-volume") pod "7227e37e-faca-45f5-ba13-0877ea8fc688" (UID: "7227e37e-faca-45f5-ba13-0877ea8fc688"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 16:30:02 crc kubenswrapper[4936]: I0320 16:30:02.983363 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7227e37e-faca-45f5-ba13-0877ea8fc688-kube-api-access-wvnxt" (OuterVolumeSpecName: "kube-api-access-wvnxt") pod "7227e37e-faca-45f5-ba13-0877ea8fc688" (UID: "7227e37e-faca-45f5-ba13-0877ea8fc688"). InnerVolumeSpecName "kube-api-access-wvnxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:30:02 crc kubenswrapper[4936]: I0320 16:30:02.984025 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7227e37e-faca-45f5-ba13-0877ea8fc688-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7227e37e-faca-45f5-ba13-0877ea8fc688" (UID: "7227e37e-faca-45f5-ba13-0877ea8fc688"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 16:30:03 crc kubenswrapper[4936]: I0320 16:30:03.071513 4936 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7227e37e-faca-45f5-ba13-0877ea8fc688-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 16:30:03 crc kubenswrapper[4936]: I0320 16:30:03.071776 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvnxt\" (UniqueName: \"kubernetes.io/projected/7227e37e-faca-45f5-ba13-0877ea8fc688-kube-api-access-wvnxt\") on node \"crc\" DevicePath \"\"" Mar 20 16:30:03 crc kubenswrapper[4936]: I0320 16:30:03.071788 4936 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7227e37e-faca-45f5-ba13-0877ea8fc688-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 16:30:03 crc kubenswrapper[4936]: I0320 16:30:03.601145 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" event={"ID":"7227e37e-faca-45f5-ba13-0877ea8fc688","Type":"ContainerDied","Data":"60a1962f32d6c887d226ae0038120830d7ac705b4cce78122e0a56839829ea91"} Mar 20 16:30:03 crc kubenswrapper[4936]: I0320 16:30:03.601444 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60a1962f32d6c887d226ae0038120830d7ac705b4cce78122e0a56839829ea91" Mar 20 16:30:03 crc kubenswrapper[4936]: I0320 16:30:03.601202 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567070-252k5" Mar 20 16:30:05 crc kubenswrapper[4936]: I0320 16:30:05.624577 4936 generic.go:334] "Generic (PLEG): container finished" podID="19b8c886-feae-44d0-a169-e07619ccad1b" containerID="b1603201a188869f6815314ba83c9b6493a66bd36799076ea331186a6033bda3" exitCode=0 Mar 20 16:30:05 crc kubenswrapper[4936]: I0320 16:30:05.624715 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567070-ljgst" event={"ID":"19b8c886-feae-44d0-a169-e07619ccad1b","Type":"ContainerDied","Data":"b1603201a188869f6815314ba83c9b6493a66bd36799076ea331186a6033bda3"} Mar 20 16:30:06 crc kubenswrapper[4936]: I0320 16:30:06.966560 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567070-ljgst" Mar 20 16:30:07 crc kubenswrapper[4936]: I0320 16:30:07.052600 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-n7kq7"] Mar 20 16:30:07 crc kubenswrapper[4936]: I0320 16:30:07.100911 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-n7kq7"] Mar 20 16:30:07 crc kubenswrapper[4936]: I0320 16:30:07.143321 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7cz4\" (UniqueName: \"kubernetes.io/projected/19b8c886-feae-44d0-a169-e07619ccad1b-kube-api-access-d7cz4\") pod \"19b8c886-feae-44d0-a169-e07619ccad1b\" (UID: \"19b8c886-feae-44d0-a169-e07619ccad1b\") " Mar 20 16:30:07 crc kubenswrapper[4936]: I0320 16:30:07.148746 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19b8c886-feae-44d0-a169-e07619ccad1b-kube-api-access-d7cz4" (OuterVolumeSpecName: "kube-api-access-d7cz4") pod "19b8c886-feae-44d0-a169-e07619ccad1b" (UID: "19b8c886-feae-44d0-a169-e07619ccad1b"). InnerVolumeSpecName "kube-api-access-d7cz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:30:07 crc kubenswrapper[4936]: I0320 16:30:07.245423 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7cz4\" (UniqueName: \"kubernetes.io/projected/19b8c886-feae-44d0-a169-e07619ccad1b-kube-api-access-d7cz4\") on node \"crc\" DevicePath \"\"" Mar 20 16:30:07 crc kubenswrapper[4936]: I0320 16:30:07.653615 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567070-ljgst" event={"ID":"19b8c886-feae-44d0-a169-e07619ccad1b","Type":"ContainerDied","Data":"d3e5c3a3dc84eb9c2c8960d2bc6eda6a3d08dab98274b3ae3d1b2842a3cf5a6e"} Mar 20 16:30:07 crc kubenswrapper[4936]: I0320 16:30:07.653702 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3e5c3a3dc84eb9c2c8960d2bc6eda6a3d08dab98274b3ae3d1b2842a3cf5a6e" Mar 20 16:30:07 crc kubenswrapper[4936]: I0320 16:30:07.653726 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567070-ljgst" Mar 20 16:30:07 crc kubenswrapper[4936]: I0320 16:30:07.882942 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="047fb63e-64a8-4280-9846-659575038df3" path="/var/lib/kubelet/pods/047fb63e-64a8-4280-9846-659575038df3/volumes" Mar 20 16:30:08 crc kubenswrapper[4936]: I0320 16:30:08.039717 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567064-8kqzc"] Mar 20 16:30:08 crc kubenswrapper[4936]: I0320 16:30:08.048504 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567064-8kqzc"] Mar 20 16:30:09 crc kubenswrapper[4936]: I0320 16:30:09.869837 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83b71159-8ac2-4910-86ea-639368e9fe03" path="/var/lib/kubelet/pods/83b71159-8ac2-4910-86ea-639368e9fe03/volumes" Mar 20 16:30:28 crc kubenswrapper[4936]: I0320 16:30:28.855181 4936 patch_prober.go:28] interesting pod/machine-config-daemon-4cxh6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 16:30:28 crc kubenswrapper[4936]: I0320 16:30:28.855709 4936 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 16:30:28 crc kubenswrapper[4936]: I0320 16:30:28.855758 4936 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" Mar 20 16:30:28 crc kubenswrapper[4936]: I0320 16:30:28.856383 4936 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a"} pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 16:30:28 crc kubenswrapper[4936]: I0320 16:30:28.856447 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerName="machine-config-daemon" containerID="cri-o://718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" gracePeriod=600 Mar 20 16:30:28 crc kubenswrapper[4936]: E0320 16:30:28.978564 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:30:29 crc kubenswrapper[4936]: I0320 16:30:29.861556 4936 generic.go:334] "Generic (PLEG): container finished" podID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" exitCode=0 Mar 20 16:30:29 crc kubenswrapper[4936]: I0320 16:30:29.865999 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" event={"ID":"dc3fb53f-2e69-4e94-bfa6-762afabe9063","Type":"ContainerDied","Data":"718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a"} Mar 20 16:30:29 crc kubenswrapper[4936]: I0320 16:30:29.866174 4936 scope.go:117] "RemoveContainer" containerID="7f54448fa33df6b9490f8d0906969969ecfdaf5a9c74fdcedd6ab26b4c513d8e" Mar 20 16:30:29 crc kubenswrapper[4936]: I0320 16:30:29.866769 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:30:29 crc kubenswrapper[4936]: E0320 16:30:29.867120 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:30:42 crc kubenswrapper[4936]: I0320 16:30:42.391770 4936 scope.go:117] "RemoveContainer" containerID="5c9803b6aad8ae456f72f069496c28ed3be945dee4fe018b3c421ef016932f5e" Mar 20 16:30:42 crc kubenswrapper[4936]: I0320 16:30:42.434977 4936 scope.go:117] "RemoveContainer" containerID="f26d39a1536d4b0e64ac2186714e024431f3e1d2286b9b8bc68f44c8be5c1b3e" Mar 20 16:30:42 crc kubenswrapper[4936]: I0320 16:30:42.498534 4936 scope.go:117] "RemoveContainer" containerID="8aa8f7bd2be84c11c343b67ab7885d9bb71ba970c84a77f1047667af257cb1f9" Mar 20 16:30:42 crc kubenswrapper[4936]: I0320 16:30:42.525844 4936 scope.go:117] "RemoveContainer" containerID="198ef6b866a3a9d1b7dc2c7023f673c2fcccad5662c4896ba53b259233660520" Mar 20 16:30:42 crc kubenswrapper[4936]: I0320 16:30:42.569675 4936 scope.go:117] "RemoveContainer" containerID="7c4fc58febfccc4dbd4458de7a1dee9f847f11da75accafe6e2b829b70cd92f0" Mar 20 16:30:42 crc kubenswrapper[4936]: I0320 16:30:42.623628 4936 scope.go:117] "RemoveContainer" containerID="6ea8047ac48d471af2f2db2043810141564ff85eb8ce2ea84b1c0e7da951f117" Mar 20 16:30:42 crc kubenswrapper[4936]: I0320 16:30:42.669860 4936 scope.go:117] "RemoveContainer" containerID="35818d37873ba37611214468b71d71e9107f339eac76ba0c3a162bac351a8d84" Mar 20 16:30:43 crc kubenswrapper[4936]: I0320 16:30:43.854953 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:30:43 crc kubenswrapper[4936]: E0320 16:30:43.855261 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:30:48 crc kubenswrapper[4936]: I0320 16:30:48.043252 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-ktpms"] Mar 20 16:30:48 crc kubenswrapper[4936]: I0320 16:30:48.053958 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-ktpms"] Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.032594 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-tv2ph"] Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.041861 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-f9qdx"] Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.051764 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-06af-account-create-update-t8zxx"] Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.059641 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-f9qdx"] Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.066554 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-tv2ph"] Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.073147 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-06af-account-create-update-t8zxx"] Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.868740 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04a90ca0-4946-4286-86b1-ba337895f3e8" path="/var/lib/kubelet/pods/04a90ca0-4946-4286-86b1-ba337895f3e8/volumes" Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.869370 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9093776-e539-4e6e-ac2e-01da662cbe84" path="/var/lib/kubelet/pods/d9093776-e539-4e6e-ac2e-01da662cbe84/volumes" Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.869995 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec06ab44-bb2e-4add-8408-49d78727c812" path="/var/lib/kubelet/pods/ec06ab44-bb2e-4add-8408-49d78727c812/volumes" Mar 20 16:30:49 crc kubenswrapper[4936]: I0320 16:30:49.870492 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee8f6e3e-3bed-4aca-b7b4-98611a317a19" path="/var/lib/kubelet/pods/ee8f6e3e-3bed-4aca-b7b4-98611a317a19/volumes" Mar 20 16:30:50 crc kubenswrapper[4936]: I0320 16:30:50.037246 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-685b-account-create-update-j7zd9"] Mar 20 16:30:50 crc kubenswrapper[4936]: I0320 16:30:50.047128 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-685b-account-create-update-j7zd9"] Mar 20 16:30:51 crc kubenswrapper[4936]: I0320 16:30:51.043052 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-416f-account-create-update-dk82t"] Mar 20 16:30:51 crc kubenswrapper[4936]: I0320 16:30:51.062459 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-416f-account-create-update-dk82t"] Mar 20 16:30:51 crc kubenswrapper[4936]: I0320 16:30:51.865972 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4111fca-7b2f-456e-9591-e9f4b8f764a2" path="/var/lib/kubelet/pods/a4111fca-7b2f-456e-9591-e9f4b8f764a2/volumes" Mar 20 16:30:51 crc kubenswrapper[4936]: I0320 16:30:51.866874 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f978ca59-3511-4a0f-9ae4-944e843f1c45" path="/var/lib/kubelet/pods/f978ca59-3511-4a0f-9ae4-944e843f1c45/volumes" Mar 20 16:30:55 crc kubenswrapper[4936]: I0320 16:30:55.855155 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:30:55 crc kubenswrapper[4936]: E0320 16:30:55.855998 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:31:06 crc kubenswrapper[4936]: I0320 16:31:06.855488 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:31:06 crc kubenswrapper[4936]: E0320 16:31:06.856899 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:31:21 crc kubenswrapper[4936]: I0320 16:31:21.070323 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hspgq"] Mar 20 16:31:21 crc kubenswrapper[4936]: I0320 16:31:21.084449 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hspgq"] Mar 20 16:31:21 crc kubenswrapper[4936]: I0320 16:31:21.854435 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:31:21 crc kubenswrapper[4936]: E0320 16:31:21.854808 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:31:21 crc kubenswrapper[4936]: I0320 16:31:21.866371 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49bf6254-7f19-41f0-969b-fc8b338ece00" path="/var/lib/kubelet/pods/49bf6254-7f19-41f0-969b-fc8b338ece00/volumes" Mar 20 16:31:33 crc kubenswrapper[4936]: I0320 16:31:33.855453 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:31:33 crc kubenswrapper[4936]: E0320 16:31:33.856488 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:31:40 crc kubenswrapper[4936]: I0320 16:31:40.047501 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-k54pt"] Mar 20 16:31:40 crc kubenswrapper[4936]: I0320 16:31:40.059765 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-k54pt"] Mar 20 16:31:41 crc kubenswrapper[4936]: I0320 16:31:41.866085 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="446d671c-a5e1-4430-9aa6-262bb3a9b2d0" path="/var/lib/kubelet/pods/446d671c-a5e1-4430-9aa6-262bb3a9b2d0/volumes" Mar 20 16:31:42 crc kubenswrapper[4936]: I0320 16:31:42.840170 4936 scope.go:117] "RemoveContainer" containerID="6590eaa90b73ad5b235781e69a2f1813aa5b5a16bbd87bb0d2552ced2ad89123" Mar 20 16:31:42 crc kubenswrapper[4936]: I0320 16:31:42.866488 4936 scope.go:117] "RemoveContainer" containerID="df2185c05f090644e7cfe6108e193ef3f7821f9aa7a7ebb86bba8d2c486395aa" Mar 20 16:31:42 crc kubenswrapper[4936]: I0320 16:31:42.933172 4936 scope.go:117] "RemoveContainer" containerID="a1fda63ccf01efaf1986d551a0706a3f97e0f7ea62addcfe032a26f1d4882479" Mar 20 16:31:42 crc kubenswrapper[4936]: I0320 16:31:42.974851 4936 scope.go:117] "RemoveContainer" containerID="dd70f68a07ba25ef2d0a249aa2dd0e221a36c633e047a1871995e849f4664344" Mar 20 16:31:43 crc kubenswrapper[4936]: I0320 16:31:43.027718 4936 scope.go:117] "RemoveContainer" containerID="c18db7a3d9576592f70a01ee0599f74b09c962beacbd3cf95e3dfeeaf1d6b493" Mar 20 16:31:43 crc kubenswrapper[4936]: I0320 16:31:43.085389 4936 scope.go:117] "RemoveContainer" containerID="820e168264596bbee554652ab395adf1401820fca494d072f9a3ff01ec0d1db8" Mar 20 16:31:43 crc kubenswrapper[4936]: I0320 16:31:43.107654 4936 scope.go:117] "RemoveContainer" containerID="4eb396796326dbce4cf8a304e85aeb386499f7e4c7e8e4cdde3eab2678927f96" Mar 20 16:31:43 crc kubenswrapper[4936]: I0320 16:31:43.127981 4936 scope.go:117] "RemoveContainer" containerID="4ec4af133a8429106464a52c50ede79627d59cde86c3be8f1687b7ce31757ee7" Mar 20 16:31:47 crc kubenswrapper[4936]: I0320 16:31:47.854410 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:31:47 crc kubenswrapper[4936]: E0320 16:31:47.855421 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:31:48 crc kubenswrapper[4936]: I0320 16:31:48.031267 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fvb75"] Mar 20 16:31:48 crc kubenswrapper[4936]: I0320 16:31:48.040392 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fvb75"] Mar 20 16:31:49 crc kubenswrapper[4936]: I0320 16:31:49.884231 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6071aca2-12af-427f-88a0-5661b147f6d8" path="/var/lib/kubelet/pods/6071aca2-12af-427f-88a0-5661b147f6d8/volumes" Mar 20 16:31:59 crc kubenswrapper[4936]: I0320 16:31:59.859671 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:31:59 crc kubenswrapper[4936]: E0320 16:31:59.861780 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.165577 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567072-hn8lw"] Mar 20 16:32:00 crc kubenswrapper[4936]: E0320 16:32:00.166077 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b8c886-feae-44d0-a169-e07619ccad1b" containerName="oc" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.166099 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b8c886-feae-44d0-a169-e07619ccad1b" containerName="oc" Mar 20 16:32:00 crc kubenswrapper[4936]: E0320 16:32:00.166136 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7227e37e-faca-45f5-ba13-0877ea8fc688" containerName="collect-profiles" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.166146 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="7227e37e-faca-45f5-ba13-0877ea8fc688" containerName="collect-profiles" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.166376 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b8c886-feae-44d0-a169-e07619ccad1b" containerName="oc" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.166410 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="7227e37e-faca-45f5-ba13-0877ea8fc688" containerName="collect-profiles" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.167108 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567072-hn8lw" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.171172 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95xtt\" (UniqueName: \"kubernetes.io/projected/d6e4d72a-bac4-4040-8495-64df34cb99fd-kube-api-access-95xtt\") pod \"auto-csr-approver-29567072-hn8lw\" (UID: \"d6e4d72a-bac4-4040-8495-64df34cb99fd\") " pod="openshift-infra/auto-csr-approver-29567072-hn8lw" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.179218 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567072-hn8lw"] Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.191307 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.191657 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.192121 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.273393 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95xtt\" (UniqueName: \"kubernetes.io/projected/d6e4d72a-bac4-4040-8495-64df34cb99fd-kube-api-access-95xtt\") pod \"auto-csr-approver-29567072-hn8lw\" (UID: \"d6e4d72a-bac4-4040-8495-64df34cb99fd\") " pod="openshift-infra/auto-csr-approver-29567072-hn8lw" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.291862 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95xtt\" (UniqueName: \"kubernetes.io/projected/d6e4d72a-bac4-4040-8495-64df34cb99fd-kube-api-access-95xtt\") pod \"auto-csr-approver-29567072-hn8lw\" (UID: \"d6e4d72a-bac4-4040-8495-64df34cb99fd\") " pod="openshift-infra/auto-csr-approver-29567072-hn8lw" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.493333 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567072-hn8lw" Mar 20 16:32:00 crc kubenswrapper[4936]: I0320 16:32:00.941393 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567072-hn8lw"] Mar 20 16:32:01 crc kubenswrapper[4936]: I0320 16:32:01.730578 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567072-hn8lw" event={"ID":"d6e4d72a-bac4-4040-8495-64df34cb99fd","Type":"ContainerStarted","Data":"0b181fdd83bf53cc90a9dce441c3d256b34fb28dcfff062e463416629e0b740d"} Mar 20 16:32:02 crc kubenswrapper[4936]: I0320 16:32:02.742219 4936 generic.go:334] "Generic (PLEG): container finished" podID="d6e4d72a-bac4-4040-8495-64df34cb99fd" containerID="17aa1ea60599ffb251a726e9f394122cfc04b26bdb1bf7b218e603e730650f04" exitCode=0 Mar 20 16:32:02 crc kubenswrapper[4936]: I0320 16:32:02.742347 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567072-hn8lw" event={"ID":"d6e4d72a-bac4-4040-8495-64df34cb99fd","Type":"ContainerDied","Data":"17aa1ea60599ffb251a726e9f394122cfc04b26bdb1bf7b218e603e730650f04"} Mar 20 16:32:04 crc kubenswrapper[4936]: I0320 16:32:04.273639 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567072-hn8lw" Mar 20 16:32:04 crc kubenswrapper[4936]: I0320 16:32:04.425800 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95xtt\" (UniqueName: \"kubernetes.io/projected/d6e4d72a-bac4-4040-8495-64df34cb99fd-kube-api-access-95xtt\") pod \"d6e4d72a-bac4-4040-8495-64df34cb99fd\" (UID: \"d6e4d72a-bac4-4040-8495-64df34cb99fd\") " Mar 20 16:32:04 crc kubenswrapper[4936]: I0320 16:32:04.430690 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e4d72a-bac4-4040-8495-64df34cb99fd-kube-api-access-95xtt" (OuterVolumeSpecName: "kube-api-access-95xtt") pod "d6e4d72a-bac4-4040-8495-64df34cb99fd" (UID: "d6e4d72a-bac4-4040-8495-64df34cb99fd"). InnerVolumeSpecName "kube-api-access-95xtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:32:04 crc kubenswrapper[4936]: I0320 16:32:04.528053 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95xtt\" (UniqueName: \"kubernetes.io/projected/d6e4d72a-bac4-4040-8495-64df34cb99fd-kube-api-access-95xtt\") on node \"crc\" DevicePath \"\"" Mar 20 16:32:04 crc kubenswrapper[4936]: I0320 16:32:04.769445 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567072-hn8lw" event={"ID":"d6e4d72a-bac4-4040-8495-64df34cb99fd","Type":"ContainerDied","Data":"0b181fdd83bf53cc90a9dce441c3d256b34fb28dcfff062e463416629e0b740d"} Mar 20 16:32:04 crc kubenswrapper[4936]: I0320 16:32:04.769491 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b181fdd83bf53cc90a9dce441c3d256b34fb28dcfff062e463416629e0b740d" Mar 20 16:32:04 crc kubenswrapper[4936]: I0320 16:32:04.769499 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567072-hn8lw" Mar 20 16:32:05 crc kubenswrapper[4936]: I0320 16:32:05.351059 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567066-d4pbg"] Mar 20 16:32:05 crc kubenswrapper[4936]: I0320 16:32:05.359398 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567066-d4pbg"] Mar 20 16:32:05 crc kubenswrapper[4936]: I0320 16:32:05.872089 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70fe6af5-2344-49de-a870-913d9edffc8b" path="/var/lib/kubelet/pods/70fe6af5-2344-49de-a870-913d9edffc8b/volumes" Mar 20 16:32:12 crc kubenswrapper[4936]: I0320 16:32:12.854326 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:32:12 crc kubenswrapper[4936]: E0320 16:32:12.855228 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:32:14 crc kubenswrapper[4936]: I0320 16:32:14.883608 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vtpxj"] Mar 20 16:32:14 crc kubenswrapper[4936]: E0320 16:32:14.884482 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e4d72a-bac4-4040-8495-64df34cb99fd" containerName="oc" Mar 20 16:32:14 crc kubenswrapper[4936]: I0320 16:32:14.884497 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e4d72a-bac4-4040-8495-64df34cb99fd" containerName="oc" Mar 20 16:32:14 crc kubenswrapper[4936]: I0320 16:32:14.884707 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e4d72a-bac4-4040-8495-64df34cb99fd" containerName="oc" Mar 20 16:32:14 crc kubenswrapper[4936]: I0320 16:32:14.887599 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:14 crc kubenswrapper[4936]: I0320 16:32:14.898655 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vtpxj"] Mar 20 16:32:14 crc kubenswrapper[4936]: I0320 16:32:14.985631 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-catalog-content\") pod \"certified-operators-vtpxj\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:14 crc kubenswrapper[4936]: I0320 16:32:14.986286 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfbwq\" (UniqueName: \"kubernetes.io/projected/4f677348-78a7-46c6-82e0-321abcde8c60-kube-api-access-zfbwq\") pod \"certified-operators-vtpxj\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:14 crc kubenswrapper[4936]: I0320 16:32:14.986344 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-utilities\") pod \"certified-operators-vtpxj\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:15 crc kubenswrapper[4936]: I0320 16:32:15.088495 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-utilities\") pod \"certified-operators-vtpxj\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:15 crc kubenswrapper[4936]: I0320 16:32:15.088689 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-catalog-content\") pod \"certified-operators-vtpxj\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:15 crc kubenswrapper[4936]: I0320 16:32:15.088772 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfbwq\" (UniqueName: \"kubernetes.io/projected/4f677348-78a7-46c6-82e0-321abcde8c60-kube-api-access-zfbwq\") pod \"certified-operators-vtpxj\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:15 crc kubenswrapper[4936]: I0320 16:32:15.089467 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-catalog-content\") pod \"certified-operators-vtpxj\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:15 crc kubenswrapper[4936]: I0320 16:32:15.090155 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-utilities\") pod \"certified-operators-vtpxj\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:15 crc kubenswrapper[4936]: I0320 16:32:15.119083 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfbwq\" (UniqueName: \"kubernetes.io/projected/4f677348-78a7-46c6-82e0-321abcde8c60-kube-api-access-zfbwq\") pod \"certified-operators-vtpxj\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:15 crc kubenswrapper[4936]: I0320 16:32:15.262940 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:15 crc kubenswrapper[4936]: I0320 16:32:15.746474 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vtpxj"] Mar 20 16:32:15 crc kubenswrapper[4936]: I0320 16:32:15.867444 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtpxj" event={"ID":"4f677348-78a7-46c6-82e0-321abcde8c60","Type":"ContainerStarted","Data":"891c444ab94292eb6624f1ecdc5e6f5d5a8182cd622889d232b9a74fb1e6771c"} Mar 20 16:32:16 crc kubenswrapper[4936]: I0320 16:32:16.878836 4936 generic.go:334] "Generic (PLEG): container finished" podID="4f677348-78a7-46c6-82e0-321abcde8c60" containerID="e3a24f489c3f9fcdab6766ac29d3dec955f5834116d56ead1bf80f14175201ca" exitCode=0 Mar 20 16:32:16 crc kubenswrapper[4936]: I0320 16:32:16.878896 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtpxj" event={"ID":"4f677348-78a7-46c6-82e0-321abcde8c60","Type":"ContainerDied","Data":"e3a24f489c3f9fcdab6766ac29d3dec955f5834116d56ead1bf80f14175201ca"} Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.280127 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-69wgc"] Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.282696 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.292638 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-69wgc"] Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.339490 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4hp4\" (UniqueName: \"kubernetes.io/projected/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-kube-api-access-p4hp4\") pod \"redhat-marketplace-69wgc\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.339621 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-catalog-content\") pod \"redhat-marketplace-69wgc\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.339661 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-utilities\") pod \"redhat-marketplace-69wgc\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.441337 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4hp4\" (UniqueName: \"kubernetes.io/projected/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-kube-api-access-p4hp4\") pod \"redhat-marketplace-69wgc\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.441713 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-catalog-content\") pod \"redhat-marketplace-69wgc\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.441815 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-utilities\") pod \"redhat-marketplace-69wgc\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.442392 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-utilities\") pod \"redhat-marketplace-69wgc\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.442717 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-catalog-content\") pod \"redhat-marketplace-69wgc\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.471056 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4hp4\" (UniqueName: \"kubernetes.io/projected/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-kube-api-access-p4hp4\") pod \"redhat-marketplace-69wgc\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.602486 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:17 crc kubenswrapper[4936]: I0320 16:32:17.921852 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtpxj" event={"ID":"4f677348-78a7-46c6-82e0-321abcde8c60","Type":"ContainerStarted","Data":"a97d99b4f8e77a4b2b12d4870ddca0b01ff67c5fd8a580bf9d8c4502d09bae16"} Mar 20 16:32:18 crc kubenswrapper[4936]: I0320 16:32:18.137726 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-69wgc"] Mar 20 16:32:18 crc kubenswrapper[4936]: I0320 16:32:18.933597 4936 generic.go:334] "Generic (PLEG): container finished" podID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerID="8dd1a1fc99ad97cd61d67c57ba7ba3195672252bdb9442a5de132ca3e25c378f" exitCode=0 Mar 20 16:32:18 crc kubenswrapper[4936]: I0320 16:32:18.933668 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69wgc" event={"ID":"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d","Type":"ContainerDied","Data":"8dd1a1fc99ad97cd61d67c57ba7ba3195672252bdb9442a5de132ca3e25c378f"} Mar 20 16:32:18 crc kubenswrapper[4936]: I0320 16:32:18.934823 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69wgc" event={"ID":"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d","Type":"ContainerStarted","Data":"6b288acf153e0d65e2d975874e9a423ec86480aa0f6595fe741f728297a87502"} Mar 20 16:32:18 crc kubenswrapper[4936]: I0320 16:32:18.939842 4936 generic.go:334] "Generic (PLEG): container finished" podID="4f677348-78a7-46c6-82e0-321abcde8c60" containerID="a97d99b4f8e77a4b2b12d4870ddca0b01ff67c5fd8a580bf9d8c4502d09bae16" exitCode=0 Mar 20 16:32:18 crc kubenswrapper[4936]: I0320 16:32:18.939879 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtpxj" event={"ID":"4f677348-78a7-46c6-82e0-321abcde8c60","Type":"ContainerDied","Data":"a97d99b4f8e77a4b2b12d4870ddca0b01ff67c5fd8a580bf9d8c4502d09bae16"} Mar 20 16:32:19 crc kubenswrapper[4936]: I0320 16:32:19.956982 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69wgc" event={"ID":"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d","Type":"ContainerStarted","Data":"09ca084499fa96881b758f34bca2644de05fc7115e876dcbe811e8748eb7db0c"} Mar 20 16:32:20 crc kubenswrapper[4936]: I0320 16:32:20.965505 4936 generic.go:334] "Generic (PLEG): container finished" podID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerID="09ca084499fa96881b758f34bca2644de05fc7115e876dcbe811e8748eb7db0c" exitCode=0 Mar 20 16:32:20 crc kubenswrapper[4936]: I0320 16:32:20.965581 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69wgc" event={"ID":"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d","Type":"ContainerDied","Data":"09ca084499fa96881b758f34bca2644de05fc7115e876dcbe811e8748eb7db0c"} Mar 20 16:32:20 crc kubenswrapper[4936]: I0320 16:32:20.968005 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtpxj" event={"ID":"4f677348-78a7-46c6-82e0-321abcde8c60","Type":"ContainerStarted","Data":"be4dbeeafada105612e50ae6dfcdd0f6b92a9228dd9b3a82e85b238ea9f11c2b"} Mar 20 16:32:21 crc kubenswrapper[4936]: I0320 16:32:21.011265 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vtpxj" podStartSLOduration=4.0130698 podStartE2EDuration="7.011248166s" podCreationTimestamp="2026-03-20 16:32:14 +0000 UTC" firstStartedPulling="2026-03-20 16:32:16.880786912 +0000 UTC m=+1887.827154727" lastFinishedPulling="2026-03-20 16:32:19.878965268 +0000 UTC m=+1890.825333093" observedRunningTime="2026-03-20 16:32:21.005838561 +0000 UTC m=+1891.952206376" watchObservedRunningTime="2026-03-20 16:32:21.011248166 +0000 UTC m=+1891.957615981" Mar 20 16:32:21 crc kubenswrapper[4936]: I0320 16:32:21.979892 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69wgc" event={"ID":"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d","Type":"ContainerStarted","Data":"1b561ae7997baa5d221ba177608a9950d72835a593296015845a4e79d755f158"} Mar 20 16:32:22 crc kubenswrapper[4936]: I0320 16:32:22.001637 4936 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-69wgc" podStartSLOduration=2.394267876 podStartE2EDuration="5.001613959s" podCreationTimestamp="2026-03-20 16:32:17 +0000 UTC" firstStartedPulling="2026-03-20 16:32:18.935124807 +0000 UTC m=+1889.881492622" lastFinishedPulling="2026-03-20 16:32:21.54247089 +0000 UTC m=+1892.488838705" observedRunningTime="2026-03-20 16:32:21.998062288 +0000 UTC m=+1892.944430103" watchObservedRunningTime="2026-03-20 16:32:22.001613959 +0000 UTC m=+1892.947981774" Mar 20 16:32:25 crc kubenswrapper[4936]: I0320 16:32:25.263409 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:25 crc kubenswrapper[4936]: I0320 16:32:25.264126 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:25 crc kubenswrapper[4936]: I0320 16:32:25.342521 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:26 crc kubenswrapper[4936]: I0320 16:32:26.061935 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-snhvn"] Mar 20 16:32:26 crc kubenswrapper[4936]: I0320 16:32:26.073345 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-snhvn"] Mar 20 16:32:26 crc kubenswrapper[4936]: I0320 16:32:26.094841 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:27 crc kubenswrapper[4936]: I0320 16:32:27.603815 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:27 crc kubenswrapper[4936]: I0320 16:32:27.604064 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:27 crc kubenswrapper[4936]: I0320 16:32:27.653195 4936 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:27 crc kubenswrapper[4936]: I0320 16:32:27.854899 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:32:27 crc kubenswrapper[4936]: E0320 16:32:27.855534 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:32:27 crc kubenswrapper[4936]: I0320 16:32:27.882731 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c" path="/var/lib/kubelet/pods/3c1fc903-c2f4-44f0-9fa1-5f01b8e9972c/volumes" Mar 20 16:32:27 crc kubenswrapper[4936]: I0320 16:32:27.891436 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vtpxj"] Mar 20 16:32:28 crc kubenswrapper[4936]: I0320 16:32:28.146389 4936 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:29 crc kubenswrapper[4936]: I0320 16:32:29.044384 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vtpxj" podUID="4f677348-78a7-46c6-82e0-321abcde8c60" containerName="registry-server" containerID="cri-o://be4dbeeafada105612e50ae6dfcdd0f6b92a9228dd9b3a82e85b238ea9f11c2b" gracePeriod=2 Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.056152 4936 generic.go:334] "Generic (PLEG): container finished" podID="4f677348-78a7-46c6-82e0-321abcde8c60" containerID="be4dbeeafada105612e50ae6dfcdd0f6b92a9228dd9b3a82e85b238ea9f11c2b" exitCode=0 Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.056381 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtpxj" event={"ID":"4f677348-78a7-46c6-82e0-321abcde8c60","Type":"ContainerDied","Data":"be4dbeeafada105612e50ae6dfcdd0f6b92a9228dd9b3a82e85b238ea9f11c2b"} Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.212877 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.308286 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-utilities\") pod \"4f677348-78a7-46c6-82e0-321abcde8c60\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.309572 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfbwq\" (UniqueName: \"kubernetes.io/projected/4f677348-78a7-46c6-82e0-321abcde8c60-kube-api-access-zfbwq\") pod \"4f677348-78a7-46c6-82e0-321abcde8c60\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.309732 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-catalog-content\") pod \"4f677348-78a7-46c6-82e0-321abcde8c60\" (UID: \"4f677348-78a7-46c6-82e0-321abcde8c60\") " Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.309133 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-utilities" (OuterVolumeSpecName: "utilities") pod "4f677348-78a7-46c6-82e0-321abcde8c60" (UID: "4f677348-78a7-46c6-82e0-321abcde8c60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.311867 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.323467 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f677348-78a7-46c6-82e0-321abcde8c60-kube-api-access-zfbwq" (OuterVolumeSpecName: "kube-api-access-zfbwq") pod "4f677348-78a7-46c6-82e0-321abcde8c60" (UID: "4f677348-78a7-46c6-82e0-321abcde8c60"). InnerVolumeSpecName "kube-api-access-zfbwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.375417 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f677348-78a7-46c6-82e0-321abcde8c60" (UID: "4f677348-78a7-46c6-82e0-321abcde8c60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.413105 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfbwq\" (UniqueName: \"kubernetes.io/projected/4f677348-78a7-46c6-82e0-321abcde8c60-kube-api-access-zfbwq\") on node \"crc\" DevicePath \"\"" Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.413138 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f677348-78a7-46c6-82e0-321abcde8c60-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.483560 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-69wgc"] Mar 20 16:32:30 crc kubenswrapper[4936]: I0320 16:32:30.483789 4936 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-69wgc" podUID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerName="registry-server" containerID="cri-o://1b561ae7997baa5d221ba177608a9950d72835a593296015845a4e79d755f158" gracePeriod=2 Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.076342 4936 generic.go:334] "Generic (PLEG): container finished" podID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerID="1b561ae7997baa5d221ba177608a9950d72835a593296015845a4e79d755f158" exitCode=0 Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.076459 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69wgc" event={"ID":"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d","Type":"ContainerDied","Data":"1b561ae7997baa5d221ba177608a9950d72835a593296015845a4e79d755f158"} Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.081137 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtpxj" event={"ID":"4f677348-78a7-46c6-82e0-321abcde8c60","Type":"ContainerDied","Data":"891c444ab94292eb6624f1ecdc5e6f5d5a8182cd622889d232b9a74fb1e6771c"} Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.081210 4936 scope.go:117] "RemoveContainer" containerID="be4dbeeafada105612e50ae6dfcdd0f6b92a9228dd9b3a82e85b238ea9f11c2b" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.081431 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vtpxj" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.125268 4936 scope.go:117] "RemoveContainer" containerID="a97d99b4f8e77a4b2b12d4870ddca0b01ff67c5fd8a580bf9d8c4502d09bae16" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.127702 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vtpxj"] Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.134897 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vtpxj"] Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.147398 4936 scope.go:117] "RemoveContainer" containerID="e3a24f489c3f9fcdab6766ac29d3dec955f5834116d56ead1bf80f14175201ca" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.438120 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.531037 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-utilities\") pod \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.531192 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-catalog-content\") pod \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.531249 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4hp4\" (UniqueName: \"kubernetes.io/projected/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-kube-api-access-p4hp4\") pod \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\" (UID: \"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d\") " Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.533437 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-utilities" (OuterVolumeSpecName: "utilities") pod "c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" (UID: "c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.536815 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-kube-api-access-p4hp4" (OuterVolumeSpecName: "kube-api-access-p4hp4") pod "c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" (UID: "c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d"). InnerVolumeSpecName "kube-api-access-p4hp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.560144 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" (UID: "c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.632947 4936 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.632984 4936 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.633000 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4hp4\" (UniqueName: \"kubernetes.io/projected/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d-kube-api-access-p4hp4\") on node \"crc\" DevicePath \"\"" Mar 20 16:32:31 crc kubenswrapper[4936]: I0320 16:32:31.874390 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f677348-78a7-46c6-82e0-321abcde8c60" path="/var/lib/kubelet/pods/4f677348-78a7-46c6-82e0-321abcde8c60/volumes" Mar 20 16:32:32 crc kubenswrapper[4936]: I0320 16:32:32.090577 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69wgc" event={"ID":"c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d","Type":"ContainerDied","Data":"6b288acf153e0d65e2d975874e9a423ec86480aa0f6595fe741f728297a87502"} Mar 20 16:32:32 crc kubenswrapper[4936]: I0320 16:32:32.090895 4936 scope.go:117] "RemoveContainer" containerID="1b561ae7997baa5d221ba177608a9950d72835a593296015845a4e79d755f158" Mar 20 16:32:32 crc kubenswrapper[4936]: I0320 16:32:32.091020 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69wgc" Mar 20 16:32:32 crc kubenswrapper[4936]: I0320 16:32:32.121466 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-69wgc"] Mar 20 16:32:32 crc kubenswrapper[4936]: I0320 16:32:32.133276 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-69wgc"] Mar 20 16:32:32 crc kubenswrapper[4936]: I0320 16:32:32.134897 4936 scope.go:117] "RemoveContainer" containerID="09ca084499fa96881b758f34bca2644de05fc7115e876dcbe811e8748eb7db0c" Mar 20 16:32:32 crc kubenswrapper[4936]: I0320 16:32:32.157214 4936 scope.go:117] "RemoveContainer" containerID="8dd1a1fc99ad97cd61d67c57ba7ba3195672252bdb9442a5de132ca3e25c378f" Mar 20 16:32:33 crc kubenswrapper[4936]: I0320 16:32:33.864790 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" path="/var/lib/kubelet/pods/c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d/volumes" Mar 20 16:32:39 crc kubenswrapper[4936]: I0320 16:32:39.859511 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:32:39 crc kubenswrapper[4936]: E0320 16:32:39.860464 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:32:43 crc kubenswrapper[4936]: I0320 16:32:43.304717 4936 scope.go:117] "RemoveContainer" containerID="d07e02914919c0f0a57dcf0f6adfb03c41bf1841fe9fa0a3c9bdbc2026c6cecc" Mar 20 16:32:43 crc kubenswrapper[4936]: I0320 16:32:43.385198 4936 scope.go:117] "RemoveContainer" containerID="3ae29d5612c535bdbba0cdff24bd5f0a1c04d915d4d35c321b334d64d27e1ae5" Mar 20 16:32:43 crc kubenswrapper[4936]: I0320 16:32:43.430790 4936 scope.go:117] "RemoveContainer" containerID="38f4cd7ab4f178d8e5ada3d9e6533d3eda21750895f91e2443e3de6b0efd684a" Mar 20 16:32:52 crc kubenswrapper[4936]: I0320 16:32:52.854678 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:32:52 crc kubenswrapper[4936]: E0320 16:32:52.855456 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:33:04 crc kubenswrapper[4936]: I0320 16:33:04.854212 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:33:04 crc kubenswrapper[4936]: E0320 16:33:04.855230 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:33:15 crc kubenswrapper[4936]: I0320 16:33:15.854481 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:33:15 crc kubenswrapper[4936]: E0320 16:33:15.855707 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:33:26 crc kubenswrapper[4936]: I0320 16:33:26.853892 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:33:26 crc kubenswrapper[4936]: E0320 16:33:26.854796 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:33:38 crc kubenswrapper[4936]: I0320 16:33:38.855047 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:33:38 crc kubenswrapper[4936]: E0320 16:33:38.855772 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:33:50 crc kubenswrapper[4936]: I0320 16:33:50.853771 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:33:50 crc kubenswrapper[4936]: E0320 16:33:50.854581 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.160024 4936 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567074-gw6tp"] Mar 20 16:34:00 crc kubenswrapper[4936]: E0320 16:34:00.161338 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f677348-78a7-46c6-82e0-321abcde8c60" containerName="registry-server" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.161352 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f677348-78a7-46c6-82e0-321abcde8c60" containerName="registry-server" Mar 20 16:34:00 crc kubenswrapper[4936]: E0320 16:34:00.161374 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerName="extract-utilities" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.161380 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerName="extract-utilities" Mar 20 16:34:00 crc kubenswrapper[4936]: E0320 16:34:00.161403 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerName="extract-content" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.161408 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerName="extract-content" Mar 20 16:34:00 crc kubenswrapper[4936]: E0320 16:34:00.161416 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f677348-78a7-46c6-82e0-321abcde8c60" containerName="extract-content" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.161421 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f677348-78a7-46c6-82e0-321abcde8c60" containerName="extract-content" Mar 20 16:34:00 crc kubenswrapper[4936]: E0320 16:34:00.161433 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f677348-78a7-46c6-82e0-321abcde8c60" containerName="extract-utilities" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.161439 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f677348-78a7-46c6-82e0-321abcde8c60" containerName="extract-utilities" Mar 20 16:34:00 crc kubenswrapper[4936]: E0320 16:34:00.161450 4936 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerName="registry-server" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.161457 4936 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerName="registry-server" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.161681 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f677348-78a7-46c6-82e0-321abcde8c60" containerName="registry-server" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.161715 4936 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b793c8-b19d-44b5-a3dc-cbf9c02f0f4d" containerName="registry-server" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.162387 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567074-gw6tp" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.165325 4936 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-sh2h6" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.165501 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.165933 4936 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.175212 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567074-gw6tp"] Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.334445 4936 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2698\" (UniqueName: \"kubernetes.io/projected/637dff5c-9977-4414-bc80-a871c6021a49-kube-api-access-l2698\") pod \"auto-csr-approver-29567074-gw6tp\" (UID: \"637dff5c-9977-4414-bc80-a871c6021a49\") " pod="openshift-infra/auto-csr-approver-29567074-gw6tp" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.435797 4936 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2698\" (UniqueName: \"kubernetes.io/projected/637dff5c-9977-4414-bc80-a871c6021a49-kube-api-access-l2698\") pod \"auto-csr-approver-29567074-gw6tp\" (UID: \"637dff5c-9977-4414-bc80-a871c6021a49\") " pod="openshift-infra/auto-csr-approver-29567074-gw6tp" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.455148 4936 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2698\" (UniqueName: \"kubernetes.io/projected/637dff5c-9977-4414-bc80-a871c6021a49-kube-api-access-l2698\") pod \"auto-csr-approver-29567074-gw6tp\" (UID: \"637dff5c-9977-4414-bc80-a871c6021a49\") " pod="openshift-infra/auto-csr-approver-29567074-gw6tp" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.486681 4936 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567074-gw6tp" Mar 20 16:34:00 crc kubenswrapper[4936]: I0320 16:34:00.918914 4936 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567074-gw6tp"] Mar 20 16:34:01 crc kubenswrapper[4936]: I0320 16:34:01.883584 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567074-gw6tp" event={"ID":"637dff5c-9977-4414-bc80-a871c6021a49","Type":"ContainerStarted","Data":"b86b2fb8f974fecd0c0bbc2dcc6d18a5164662eeb75fcb5bc40032cb4146c55e"} Mar 20 16:34:02 crc kubenswrapper[4936]: I0320 16:34:02.894666 4936 generic.go:334] "Generic (PLEG): container finished" podID="637dff5c-9977-4414-bc80-a871c6021a49" containerID="e7e00bd5b618b982ecef021eb8bb4cbcf59022e502b09b75c18b662e2d93831c" exitCode=0 Mar 20 16:34:02 crc kubenswrapper[4936]: I0320 16:34:02.894913 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567074-gw6tp" event={"ID":"637dff5c-9977-4414-bc80-a871c6021a49","Type":"ContainerDied","Data":"e7e00bd5b618b982ecef021eb8bb4cbcf59022e502b09b75c18b662e2d93831c"} Mar 20 16:34:04 crc kubenswrapper[4936]: I0320 16:34:04.263678 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567074-gw6tp" Mar 20 16:34:04 crc kubenswrapper[4936]: I0320 16:34:04.412278 4936 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2698\" (UniqueName: \"kubernetes.io/projected/637dff5c-9977-4414-bc80-a871c6021a49-kube-api-access-l2698\") pod \"637dff5c-9977-4414-bc80-a871c6021a49\" (UID: \"637dff5c-9977-4414-bc80-a871c6021a49\") " Mar 20 16:34:04 crc kubenswrapper[4936]: I0320 16:34:04.426366 4936 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/637dff5c-9977-4414-bc80-a871c6021a49-kube-api-access-l2698" (OuterVolumeSpecName: "kube-api-access-l2698") pod "637dff5c-9977-4414-bc80-a871c6021a49" (UID: "637dff5c-9977-4414-bc80-a871c6021a49"). InnerVolumeSpecName "kube-api-access-l2698". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 16:34:04 crc kubenswrapper[4936]: I0320 16:34:04.514629 4936 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2698\" (UniqueName: \"kubernetes.io/projected/637dff5c-9977-4414-bc80-a871c6021a49-kube-api-access-l2698\") on node \"crc\" DevicePath \"\"" Mar 20 16:34:04 crc kubenswrapper[4936]: I0320 16:34:04.923264 4936 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567074-gw6tp" event={"ID":"637dff5c-9977-4414-bc80-a871c6021a49","Type":"ContainerDied","Data":"b86b2fb8f974fecd0c0bbc2dcc6d18a5164662eeb75fcb5bc40032cb4146c55e"} Mar 20 16:34:04 crc kubenswrapper[4936]: I0320 16:34:04.923345 4936 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b86b2fb8f974fecd0c0bbc2dcc6d18a5164662eeb75fcb5bc40032cb4146c55e" Mar 20 16:34:04 crc kubenswrapper[4936]: I0320 16:34:04.923306 4936 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567074-gw6tp" Mar 20 16:34:05 crc kubenswrapper[4936]: I0320 16:34:05.336736 4936 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567068-vfxgs"] Mar 20 16:34:05 crc kubenswrapper[4936]: I0320 16:34:05.345109 4936 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567068-vfxgs"] Mar 20 16:34:05 crc kubenswrapper[4936]: I0320 16:34:05.854340 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:34:05 crc kubenswrapper[4936]: E0320 16:34:05.854770 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" Mar 20 16:34:05 crc kubenswrapper[4936]: I0320 16:34:05.864434 4936 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c5319ea-6340-41c9-ae05-0ecfed359126" path="/var/lib/kubelet/pods/2c5319ea-6340-41c9-ae05-0ecfed359126/volumes" Mar 20 16:34:18 crc kubenswrapper[4936]: I0320 16:34:18.854714 4936 scope.go:117] "RemoveContainer" containerID="718ad9f162c5317822e445800e642d2883376328a423da78209da98af2ee5b3a" Mar 20 16:34:18 crc kubenswrapper[4936]: E0320 16:34:18.855327 4936 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4cxh6_openshift-machine-config-operator(dc3fb53f-2e69-4e94-bfa6-762afabe9063)\"" pod="openshift-machine-config-operator/machine-config-daemon-4cxh6" podUID="dc3fb53f-2e69-4e94-bfa6-762afabe9063" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515157273433024457 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015157273433017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015157267143016520 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015157267143015470 5ustar corecore